Picture for Nicolas Perrin

Nicolas Perrin

Learning Compositional Neural Programs for Continuous Control

Add code
Jul 27, 2020
Figure 1 for Learning Compositional Neural Programs for Continuous Control
Figure 2 for Learning Compositional Neural Programs for Continuous Control
Figure 3 for Learning Compositional Neural Programs for Continuous Control
Figure 4 for Learning Compositional Neural Programs for Continuous Control
Viaarxiv icon

QD-RL: Efficient Mixing of Quality and Diversity in Reinforcement Learning

Add code
Jun 15, 2020
Figure 1 for QD-RL: Efficient Mixing of Quality and Diversity in Reinforcement Learning
Figure 2 for QD-RL: Efficient Mixing of Quality and Diversity in Reinforcement Learning
Figure 3 for QD-RL: Efficient Mixing of Quality and Diversity in Reinforcement Learning
Figure 4 for QD-RL: Efficient Mixing of Quality and Diversity in Reinforcement Learning
Viaarxiv icon

Recurrent Neural Networks for Stochastic Control in Real-Time Bidding

Add code
Jun 12, 2020
Figure 1 for Recurrent Neural Networks for Stochastic Control in Real-Time Bidding
Figure 2 for Recurrent Neural Networks for Stochastic Control in Real-Time Bidding
Figure 3 for Recurrent Neural Networks for Stochastic Control in Real-Time Bidding
Figure 4 for Recurrent Neural Networks for Stochastic Control in Real-Time Bidding
Viaarxiv icon

PBCS : Efficient Exploration and Exploitation Using a Synergy between Reinforcement Learning and Motion Planning

Add code
Apr 24, 2020
Figure 1 for PBCS : Efficient Exploration and Exploitation Using a Synergy between Reinforcement Learning and Motion Planning
Figure 2 for PBCS : Efficient Exploration and Exploitation Using a Synergy between Reinforcement Learning and Motion Planning
Viaarxiv icon

The problem with DDPG: understanding failures in deterministic environments with sparse rewards

Add code
Nov 26, 2019
Figure 1 for The problem with DDPG: understanding failures in deterministic environments with sparse rewards
Figure 2 for The problem with DDPG: understanding failures in deterministic environments with sparse rewards
Figure 3 for The problem with DDPG: understanding failures in deterministic environments with sparse rewards
Figure 4 for The problem with DDPG: understanding failures in deterministic environments with sparse rewards
Viaarxiv icon

State Representation Learning from Demonstration

Add code
Sep 15, 2019
Figure 1 for State Representation Learning from Demonstration
Figure 2 for State Representation Learning from Demonstration
Figure 3 for State Representation Learning from Demonstration
Figure 4 for State Representation Learning from Demonstration
Viaarxiv icon

Learning Compositional Neural Programs with Recursive Tree Search and Planning

Add code
May 30, 2019
Figure 1 for Learning Compositional Neural Programs with Recursive Tree Search and Planning
Figure 2 for Learning Compositional Neural Programs with Recursive Tree Search and Planning
Figure 3 for Learning Compositional Neural Programs with Recursive Tree Search and Planning
Figure 4 for Learning Compositional Neural Programs with Recursive Tree Search and Planning
Viaarxiv icon

First-order and second-order variants of the gradient descent: a unified framework

Add code
Oct 18, 2018
Figure 1 for First-order and second-order variants of the gradient descent: a unified framework
Figure 2 for First-order and second-order variants of the gradient descent: a unified framework
Figure 3 for First-order and second-order variants of the gradient descent: a unified framework
Viaarxiv icon

Importance mixing: Improving sample reuse in evolutionary policy search methods

Add code
Aug 17, 2018
Figure 1 for Importance mixing: Improving sample reuse in evolutionary policy search methods
Figure 2 for Importance mixing: Improving sample reuse in evolutionary policy search methods
Figure 3 for Importance mixing: Improving sample reuse in evolutionary policy search methods
Figure 4 for Importance mixing: Improving sample reuse in evolutionary policy search methods
Viaarxiv icon