Picture for Oleg Klimov

Oleg Klimov

Multi-task curriculum learning in a complex, visual, hard-exploration domain: Minecraft

Add code
Jun 28, 2021
Figure 1 for Multi-task curriculum learning in a complex, visual, hard-exploration domain: Minecraft
Figure 2 for Multi-task curriculum learning in a complex, visual, hard-exploration domain: Minecraft
Figure 3 for Multi-task curriculum learning in a complex, visual, hard-exploration domain: Minecraft
Figure 4 for Multi-task curriculum learning in a complex, visual, hard-exploration domain: Minecraft
Viaarxiv icon

Phasic Policy Gradient

Add code
Sep 09, 2020
Figure 1 for Phasic Policy Gradient
Figure 2 for Phasic Policy Gradient
Figure 3 for Phasic Policy Gradient
Figure 4 for Phasic Policy Gradient
Viaarxiv icon

Quantifying Generalization in Reinforcement Learning

Add code
Dec 20, 2018
Figure 1 for Quantifying Generalization in Reinforcement Learning
Figure 2 for Quantifying Generalization in Reinforcement Learning
Figure 3 for Quantifying Generalization in Reinforcement Learning
Figure 4 for Quantifying Generalization in Reinforcement Learning
Viaarxiv icon

Exploration by Random Network Distillation

Add code
Oct 30, 2018
Figure 1 for Exploration by Random Network Distillation
Figure 2 for Exploration by Random Network Distillation
Figure 3 for Exploration by Random Network Distillation
Figure 4 for Exploration by Random Network Distillation
Viaarxiv icon

Gotta Learn Fast: A New Benchmark for Generalization in RL

Add code
Apr 23, 2018
Figure 1 for Gotta Learn Fast: A New Benchmark for Generalization in RL
Figure 2 for Gotta Learn Fast: A New Benchmark for Generalization in RL
Figure 3 for Gotta Learn Fast: A New Benchmark for Generalization in RL
Figure 4 for Gotta Learn Fast: A New Benchmark for Generalization in RL
Viaarxiv icon

Proximal Policy Optimization Algorithms

Add code
Aug 28, 2017
Figure 1 for Proximal Policy Optimization Algorithms
Figure 2 for Proximal Policy Optimization Algorithms
Figure 3 for Proximal Policy Optimization Algorithms
Figure 4 for Proximal Policy Optimization Algorithms
Viaarxiv icon