Picture for Michal Nauman

Michal Nauman

Bigger, Regularized, Optimistic: scaling for compute and sample-efficient continuous control

Add code
May 25, 2024
Viaarxiv icon

A Case for Validation Buffer in Pessimistic Actor-Critic

Add code
Mar 01, 2024
Viaarxiv icon

Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning

Add code
Mar 01, 2024
Figure 1 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 2 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 3 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Figure 4 for Overestimation, Overfitting, and Plasticity in Actor-Critic: the Bitter Lesson of Reinforcement Learning
Viaarxiv icon

Decoupled Actor-Critic

Add code
Oct 30, 2023
Figure 1 for Decoupled Actor-Critic
Figure 2 for Decoupled Actor-Critic
Figure 3 for Decoupled Actor-Critic
Figure 4 for Decoupled Actor-Critic
Viaarxiv icon

On All-Action Policy Gradients

Add code
Oct 24, 2022
Viaarxiv icon

Low-Variance Policy Gradient Estimation with World Models

Add code
Oct 29, 2020
Figure 1 for Low-Variance Policy Gradient Estimation with World Models
Figure 2 for Low-Variance Policy Gradient Estimation with World Models
Figure 3 for Low-Variance Policy Gradient Estimation with World Models
Figure 4 for Low-Variance Policy Gradient Estimation with World Models
Viaarxiv icon