Picture for Voot Tangkaratt

Voot Tangkaratt

Discovering Diverse Solutions in Deep Reinforcement Learning

Add code
Mar 12, 2021
Figure 1 for Discovering Diverse Solutions in Deep Reinforcement Learning
Figure 2 for Discovering Diverse Solutions in Deep Reinforcement Learning
Figure 3 for Discovering Diverse Solutions in Deep Reinforcement Learning
Figure 4 for Discovering Diverse Solutions in Deep Reinforcement Learning
Viaarxiv icon

Robust Imitation Learning from Noisy Demonstrations

Add code
Oct 31, 2020
Figure 1 for Robust Imitation Learning from Noisy Demonstrations
Figure 2 for Robust Imitation Learning from Noisy Demonstrations
Figure 3 for Robust Imitation Learning from Noisy Demonstrations
Figure 4 for Robust Imitation Learning from Noisy Demonstrations
Viaarxiv icon

Meta-Model-Based Meta-Policy Optimization

Add code
Jun 05, 2020
Figure 1 for Meta-Model-Based Meta-Policy Optimization
Figure 2 for Meta-Model-Based Meta-Policy Optimization
Figure 3 for Meta-Model-Based Meta-Policy Optimization
Figure 4 for Meta-Model-Based Meta-Policy Optimization
Viaarxiv icon

VILD: Variational Imitation Learning with Diverse-quality Demonstrations

Add code
Sep 15, 2019
Figure 1 for VILD: Variational Imitation Learning with Diverse-quality Demonstrations
Figure 2 for VILD: Variational Imitation Learning with Diverse-quality Demonstrations
Figure 3 for VILD: Variational Imitation Learning with Diverse-quality Demonstrations
Figure 4 for VILD: Variational Imitation Learning with Diverse-quality Demonstrations
Viaarxiv icon

Imitation Learning from Imperfect Demonstration

Add code
Jan 30, 2019
Figure 1 for Imitation Learning from Imperfect Demonstration
Figure 2 for Imitation Learning from Imperfect Demonstration
Figure 3 for Imitation Learning from Imperfect Demonstration
Figure 4 for Imitation Learning from Imperfect Demonstration
Viaarxiv icon

Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization

Add code
Jan 05, 2019
Figure 1 for Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization
Figure 2 for Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization
Figure 3 for Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization
Figure 4 for Hierarchical Reinforcement Learning via Advantage-Weighted Information Maximization
Viaarxiv icon

TD-Regularized Actor-Critic Methods

Add code
Dec 23, 2018
Figure 1 for TD-Regularized Actor-Critic Methods
Figure 2 for TD-Regularized Actor-Critic Methods
Figure 3 for TD-Regularized Actor-Critic Methods
Figure 4 for TD-Regularized Actor-Critic Methods
Viaarxiv icon

Active Deep Q-learning with Demonstration

Add code
Dec 06, 2018
Figure 1 for Active Deep Q-learning with Demonstration
Figure 2 for Active Deep Q-learning with Demonstration
Figure 3 for Active Deep Q-learning with Demonstration
Figure 4 for Active Deep Q-learning with Demonstration
Viaarxiv icon

Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam

Add code
Aug 02, 2018
Figure 1 for Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam
Figure 2 for Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam
Figure 3 for Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam
Figure 4 for Fast and Scalable Bayesian Deep Learning by Weight-Perturbation in Adam
Viaarxiv icon

Guide Actor-Critic for Continuous Control

Add code
Feb 22, 2018
Figure 1 for Guide Actor-Critic for Continuous Control
Figure 2 for Guide Actor-Critic for Continuous Control
Figure 3 for Guide Actor-Critic for Continuous Control
Figure 4 for Guide Actor-Critic for Continuous Control
Viaarxiv icon