Picture for J. Andrew Bagnell

J. Andrew Bagnell

Hybrid Reinforcement Learning from Offline Observation Alone

Add code
Jun 11, 2024
Viaarxiv icon

Understanding Preference Fine-Tuning Through the Lens of Coverage

Add code
Jun 03, 2024
Viaarxiv icon

REBEL: Reinforcement Learning via Regressing Relative Rewards

Add code
Apr 25, 2024
Viaarxiv icon

Hybrid Inverse Reinforcement Learning

Add code
Feb 13, 2024
Viaarxiv icon

The Virtues of Pessimism in Inverse Reinforcement Learning

Add code
Feb 08, 2024
Viaarxiv icon

Inverse Reinforcement Learning without Reinforcement Learning

Add code
Mar 26, 2023
Viaarxiv icon

The Virtues of Laziness in Model-based RL: A Unified Objective and Algorithms

Add code
Mar 01, 2023
Viaarxiv icon

Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient

Add code
Oct 13, 2022
Figure 1 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Figure 2 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Figure 3 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Figure 4 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Viaarxiv icon

Game-Theoretic Algorithms for Conditional Moment Matching

Add code
Aug 19, 2022
Figure 1 for Game-Theoretic Algorithms for Conditional Moment Matching
Figure 2 for Game-Theoretic Algorithms for Conditional Moment Matching
Viaarxiv icon

Sequence Model Imitation Learning with Unobserved Contexts

Add code
Aug 03, 2022
Figure 1 for Sequence Model Imitation Learning with Unobserved Contexts
Figure 2 for Sequence Model Imitation Learning with Unobserved Contexts
Figure 3 for Sequence Model Imitation Learning with Unobserved Contexts
Figure 4 for Sequence Model Imitation Learning with Unobserved Contexts
Viaarxiv icon