Picture for Aaron Defazio

Aaron Defazio

Alice

The Road Less Scheduled

Add code
May 24, 2024
Viaarxiv icon

Directional Smoothness and Gradient Methods: Convergence and Adaptivity

Add code
Mar 06, 2024
Figure 1 for Directional Smoothness and Gradient Methods: Convergence and Adaptivity
Figure 2 for Directional Smoothness and Gradient Methods: Convergence and Adaptivity
Figure 3 for Directional Smoothness and Gradient Methods: Convergence and Adaptivity
Figure 4 for Directional Smoothness and Gradient Methods: Convergence and Adaptivity
Viaarxiv icon

When, Why and How Much? Adaptive Learning Rate Scheduling by Refinement

Add code
Oct 11, 2023
Viaarxiv icon

Prodigy: An Expeditiously Adaptive Parameter-Free Learner

Add code
Jun 09, 2023
Viaarxiv icon

Mechanic: A Learning Rate Tuner

Add code
Jun 02, 2023
Viaarxiv icon

MoMo: Momentum Models for Adaptive Learning Rates

Add code
May 12, 2023
Figure 1 for MoMo: Momentum Models for Adaptive Learning Rates
Figure 2 for MoMo: Momentum Models for Adaptive Learning Rates
Figure 3 for MoMo: Momentum Models for Adaptive Learning Rates
Figure 4 for MoMo: Momentum Models for Adaptive Learning Rates
Viaarxiv icon

Learning-Rate-Free Learning by D-Adaptation

Add code
Jan 20, 2023
Viaarxiv icon

Grad-GradaGrad? A Non-Monotone Adaptive Stochastic Gradient Method

Add code
Jun 14, 2022
Figure 1 for Grad-GradaGrad? A Non-Monotone Adaptive Stochastic Gradient Method
Figure 2 for Grad-GradaGrad? A Non-Monotone Adaptive Stochastic Gradient Method
Figure 3 for Grad-GradaGrad? A Non-Monotone Adaptive Stochastic Gradient Method
Figure 4 for Grad-GradaGrad? A Non-Monotone Adaptive Stochastic Gradient Method
Viaarxiv icon

Stochastic Polyak Stepsize with a Moving Target

Add code
Jun 22, 2021
Figure 1 for Stochastic Polyak Stepsize with a Moving Target
Figure 2 for Stochastic Polyak Stepsize with a Moving Target
Figure 3 for Stochastic Polyak Stepsize with a Moving Target
Figure 4 for Stochastic Polyak Stepsize with a Moving Target
Viaarxiv icon

Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization

Add code
Jan 26, 2021
Figure 1 for Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization
Figure 2 for Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization
Figure 3 for Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization
Figure 4 for Adaptivity without Compromise: A Momentumized, Adaptive, Dual Averaged Gradient Method for Stochastic Optimization
Viaarxiv icon