Picture for Sharan Vaswani

Sharan Vaswani

Towards Principled, Practical Policy Gradient for Bandits and Tabular MDPs

Add code
May 21, 2024
Viaarxiv icon

From Inverse Optimization to Feasibility to ERM

Add code
Feb 27, 2024
Viaarxiv icon

Noise-adaptive (Accelerated) Stochastic Heavy-Ball Momentum

Add code
Jan 12, 2024
Viaarxiv icon

Decision-Aware Actor-Critic with Function Approximation and Theoretical Guarantees

Add code
May 24, 2023
Viaarxiv icon

Target-based Surrogates for Stochastic Optimization

Add code
Feb 06, 2023
Viaarxiv icon

Improved Policy Optimization for Online Imitation Learning

Add code
Jul 29, 2022
Figure 1 for Improved Policy Optimization for Online Imitation Learning
Figure 2 for Improved Policy Optimization for Online Imitation Learning
Figure 3 for Improved Policy Optimization for Online Imitation Learning
Figure 4 for Improved Policy Optimization for Online Imitation Learning
Viaarxiv icon

Near-Optimal Sample Complexity Bounds for Constrained MDPs

Add code
Jun 13, 2022
Figure 1 for Near-Optimal Sample Complexity Bounds for Constrained MDPs
Figure 2 for Near-Optimal Sample Complexity Bounds for Constrained MDPs
Viaarxiv icon

Towards Painless Policy Optimization for Constrained MDPs

Add code
Apr 11, 2022
Figure 1 for Towards Painless Policy Optimization for Constrained MDPs
Figure 2 for Towards Painless Policy Optimization for Constrained MDPs
Figure 3 for Towards Painless Policy Optimization for Constrained MDPs
Figure 4 for Towards Painless Policy Optimization for Constrained MDPs
Viaarxiv icon

Towards Noise-adaptive, Problem-adaptive Stochastic Gradient Descent

Add code
Oct 21, 2021
Figure 1 for Towards Noise-adaptive, Problem-adaptive Stochastic Gradient Descent
Viaarxiv icon

A functional mirror ascent view of policy gradient methods with function approximation

Add code
Aug 12, 2021
Figure 1 for A functional mirror ascent view of policy gradient methods with function approximation
Figure 2 for A functional mirror ascent view of policy gradient methods with function approximation
Figure 3 for A functional mirror ascent view of policy gradient methods with function approximation
Figure 4 for A functional mirror ascent view of policy gradient methods with function approximation
Viaarxiv icon