Picture for Hadi Daneshmand

Hadi Daneshmand

Provable optimal transport with transformers: The essence of depth and prompt engineering

Add code
Oct 25, 2024
Viaarxiv icon

Transformers Learn Temporal Difference Methods for In-Context Reinforcement Learning

Add code
May 22, 2024
Viaarxiv icon

Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion

Add code
Oct 03, 2023
Figure 1 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Figure 2 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Figure 3 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Figure 4 for Towards Training Without Depth Limits: Batch Normalization Without Gradient Explosion
Viaarxiv icon

Transformers learn to implement preconditioned gradient descent for in-context learning

Add code
Jun 01, 2023
Viaarxiv icon

On the impact of activation and normalization in obtaining isometric embeddings at initialization

Add code
May 28, 2023
Viaarxiv icon

Efficient displacement convex optimization with particle gradient descent

Add code
Feb 09, 2023
Viaarxiv icon

Entropy Maximization with Depth: A Variational Principle for Random Neural Networks

Add code
May 25, 2022
Figure 1 for Entropy Maximization with Depth: A Variational Principle for Random Neural Networks
Viaarxiv icon

Polynomial-time sparse measure recovery

Add code
Apr 16, 2022
Figure 1 for Polynomial-time sparse measure recovery
Figure 2 for Polynomial-time sparse measure recovery
Figure 3 for Polynomial-time sparse measure recovery
Viaarxiv icon

Batch Normalization Orthogonalizes Representations in Deep Random Networks

Add code
Jun 07, 2021
Figure 1 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Figure 2 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Figure 3 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Figure 4 for Batch Normalization Orthogonalizes Representations in Deep Random Networks
Viaarxiv icon

Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization

Add code
Feb 23, 2021
Figure 1 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Figure 2 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Figure 3 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Figure 4 for Revisiting the Role of Euler Numerical Integration on Acceleration and Stability in Convex Optimization
Viaarxiv icon