Picture for Alessandro Lazaric

Alessandro Lazaric

INRIA Lille - Nord Europe

BFM-Zero: A Promptable Behavioral Foundation Model for Humanoid Control Using Unsupervised Reinforcement Learning

Add code
Nov 06, 2025
Viaarxiv icon

TD-JEPA: Latent-predictive Representations for Zero-Shot Reinforcement Learning

Add code
Oct 01, 2025
Viaarxiv icon

Zero-Shot Whole-Body Humanoid Control via Behavioral Foundation Models

Add code
Apr 15, 2025
Viaarxiv icon

Fast Adaptation with Behavioral Foundation Models

Add code
Apr 10, 2025
Viaarxiv icon

Temporal Difference Flows

Add code
Mar 12, 2025
Viaarxiv icon

System-2 Recommenders: Disentangling Utility and Engagement in Recommendation Systems via Temporal Point-Processes

Add code
May 29, 2024
Figure 1 for System-2 Recommenders: Disentangling Utility and Engagement in Recommendation Systems via Temporal Point-Processes
Figure 2 for System-2 Recommenders: Disentangling Utility and Engagement in Recommendation Systems via Temporal Point-Processes
Viaarxiv icon

Reinforcement Learning with Options and State Representation

Add code
Mar 25, 2024
Figure 1 for Reinforcement Learning with Options and State Representation
Figure 2 for Reinforcement Learning with Options and State Representation
Figure 3 for Reinforcement Learning with Options and State Representation
Figure 4 for Reinforcement Learning with Options and State Representation
Viaarxiv icon

Simple Ingredients for Offline Reinforcement Learning

Add code
Mar 19, 2024
Figure 1 for Simple Ingredients for Offline Reinforcement Learning
Figure 2 for Simple Ingredients for Offline Reinforcement Learning
Figure 3 for Simple Ingredients for Offline Reinforcement Learning
Figure 4 for Simple Ingredients for Offline Reinforcement Learning
Viaarxiv icon

Layered State Discovery for Incremental Autonomous Exploration

Add code
Feb 07, 2023
Viaarxiv icon

Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping

Add code
Jan 05, 2023
Figure 1 for Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping
Figure 2 for Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping
Figure 3 for Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping
Figure 4 for Learning Goal-Conditioned Policies Offline with Self-Supervised Reward Shaping
Viaarxiv icon