Picture for Donald Goldfarb

Donald Goldfarb

Layer-wise Adaptive Step-Sizes for Stochastic First-Order Methods for Deep Learning

Add code
May 23, 2023
Viaarxiv icon

A Mini-Block Natural Gradient Method for Deep Neural Networks

Add code
Feb 16, 2022
Figure 1 for A Mini-Block Natural Gradient Method for Deep Neural Networks
Figure 2 for A Mini-Block Natural Gradient Method for Deep Neural Networks
Figure 3 for A Mini-Block Natural Gradient Method for Deep Neural Networks
Figure 4 for A Mini-Block Natural Gradient Method for Deep Neural Networks
Viaarxiv icon

Tensor Normal Training for Deep Learning Models

Add code
Jun 05, 2021
Figure 1 for Tensor Normal Training for Deep Learning Models
Figure 2 for Tensor Normal Training for Deep Learning Models
Figure 3 for Tensor Normal Training for Deep Learning Models
Figure 4 for Tensor Normal Training for Deep Learning Models
Viaarxiv icon

Kronecker-factored Quasi-Newton Methods for Convolutional Neural Networks

Add code
Feb 12, 2021
Figure 1 for Kronecker-factored Quasi-Newton Methods for Convolutional Neural Networks
Figure 2 for Kronecker-factored Quasi-Newton Methods for Convolutional Neural Networks
Figure 3 for Kronecker-factored Quasi-Newton Methods for Convolutional Neural Networks
Figure 4 for Kronecker-factored Quasi-Newton Methods for Convolutional Neural Networks
Viaarxiv icon

Practical Quasi-Newton Methods for Training Deep Neural Networks

Add code
Jun 16, 2020
Figure 1 for Practical Quasi-Newton Methods for Training Deep Neural Networks
Figure 2 for Practical Quasi-Newton Methods for Training Deep Neural Networks
Figure 3 for Practical Quasi-Newton Methods for Training Deep Neural Networks
Figure 4 for Practical Quasi-Newton Methods for Training Deep Neural Networks
Viaarxiv icon

A Dynamic Sampling Adaptive-SGD Method for Machine Learning

Add code
Dec 31, 2019
Figure 1 for A Dynamic Sampling Adaptive-SGD Method for Machine Learning
Figure 2 for A Dynamic Sampling Adaptive-SGD Method for Machine Learning
Figure 3 for A Dynamic Sampling Adaptive-SGD Method for Machine Learning
Figure 4 for A Dynamic Sampling Adaptive-SGD Method for Machine Learning
Viaarxiv icon

Efficient Subsampled Gauss-Newton and Natural Gradient Methods for Training Neural Networks

Add code
Jun 05, 2019
Figure 1 for Efficient Subsampled Gauss-Newton and Natural Gradient Methods for Training Neural Networks
Figure 2 for Efficient Subsampled Gauss-Newton and Natural Gradient Methods for Training Neural Networks
Figure 3 for Efficient Subsampled Gauss-Newton and Natural Gradient Methods for Training Neural Networks
Viaarxiv icon

Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models

Add code
May 24, 2019
Figure 1 for Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models
Figure 2 for Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models
Figure 3 for Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models
Figure 4 for Leader Stochastic Gradient Descent for Distributed Training of Deep Learning Models
Viaarxiv icon

Scalable Robust Matrix Recovery: Frank-Wolfe Meets Proximal Methods

Add code
May 29, 2017
Figure 1 for Scalable Robust Matrix Recovery: Frank-Wolfe Meets Proximal Methods
Figure 2 for Scalable Robust Matrix Recovery: Frank-Wolfe Meets Proximal Methods
Figure 3 for Scalable Robust Matrix Recovery: Frank-Wolfe Meets Proximal Methods
Figure 4 for Scalable Robust Matrix Recovery: Frank-Wolfe Meets Proximal Methods
Viaarxiv icon

Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization

Add code
May 21, 2017
Figure 1 for Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization
Figure 2 for Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization
Figure 3 for Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization
Figure 4 for Stochastic Quasi-Newton Methods for Nonconvex Stochastic Optimization
Viaarxiv icon