Picture for Hideaki Iiduka

Hideaki Iiduka

Faster Convergence of Riemannian Stochastic Gradient Descent with Increasing Batch Size

Add code
Jan 30, 2025
Figure 1 for Faster Convergence of Riemannian Stochastic Gradient Descent with Increasing Batch Size
Figure 2 for Faster Convergence of Riemannian Stochastic Gradient Descent with Increasing Batch Size
Figure 3 for Faster Convergence of Riemannian Stochastic Gradient Descent with Increasing Batch Size
Figure 4 for Faster Convergence of Riemannian Stochastic Gradient Descent with Increasing Batch Size
Viaarxiv icon

Increasing Batch Size Improves Convergence of Stochastic Gradient Descent with Momentum

Add code
Jan 15, 2025
Viaarxiv icon

Scaled Conjugate Gradient Method for Nonconvex Optimization in Deep Neural Networks

Add code
Dec 16, 2024
Viaarxiv icon

Explicit and Implicit Graduated Optimization in Deep Neural Networks

Add code
Dec 16, 2024
Figure 1 for Explicit and Implicit Graduated Optimization in Deep Neural Networks
Figure 2 for Explicit and Implicit Graduated Optimization in Deep Neural Networks
Figure 3 for Explicit and Implicit Graduated Optimization in Deep Neural Networks
Figure 4 for Explicit and Implicit Graduated Optimization in Deep Neural Networks
Viaarxiv icon

Convergence of Sharpness-Aware Minimization Algorithms using Increasing Batch Size and Decaying Learning Rate

Add code
Sep 16, 2024
Viaarxiv icon

Increasing Both Batch Size and Learning Rate Accelerates Stochastic Gradient Descent

Add code
Sep 13, 2024
Figure 1 for Increasing Both Batch Size and Learning Rate Accelerates Stochastic Gradient Descent
Figure 2 for Increasing Both Batch Size and Learning Rate Accelerates Stochastic Gradient Descent
Figure 3 for Increasing Both Batch Size and Learning Rate Accelerates Stochastic Gradient Descent
Figure 4 for Increasing Both Batch Size and Learning Rate Accelerates Stochastic Gradient Descent
Viaarxiv icon

Iteration and Stochastic First-order Oracle Complexities of Stochastic Gradient Descent using Constant and Decaying Learning Rates

Add code
Feb 23, 2024
Figure 1 for Iteration and Stochastic First-order Oracle Complexities of Stochastic Gradient Descent using Constant and Decaying Learning Rates
Figure 2 for Iteration and Stochastic First-order Oracle Complexities of Stochastic Gradient Descent using Constant and Decaying Learning Rates
Figure 3 for Iteration and Stochastic First-order Oracle Complexities of Stochastic Gradient Descent using Constant and Decaying Learning Rates
Figure 4 for Iteration and Stochastic First-order Oracle Complexities of Stochastic Gradient Descent using Constant and Decaying Learning Rates
Viaarxiv icon

Role of Momentum in Smoothing Objective Function in Implicit Graduated Optimization

Add code
Feb 04, 2024
Figure 1 for Role of Momentum in Smoothing Objective Function in Implicit Graduated Optimization
Figure 2 for Role of Momentum in Smoothing Objective Function in Implicit Graduated Optimization
Figure 3 for Role of Momentum in Smoothing Objective Function in Implicit Graduated Optimization
Figure 4 for Role of Momentum in Smoothing Objective Function in Implicit Graduated Optimization
Viaarxiv icon

Using Stochastic Gradient Descent to Smooth Nonconvex Functions: Analysis of Implicit Graduated Optimization with Optimal Noise Scheduling

Add code
Nov 29, 2023
Viaarxiv icon

Relationship between Batch Size and Number of Steps Needed for Nonconvex Optimization of Stochastic Gradient Descent using Armijo Line Search

Add code
Aug 03, 2023
Viaarxiv icon