Picture for Yi-Chen Li

Yi-Chen Li

Controlling Large Language Model with Latent Actions

Add code
Mar 27, 2025
Viaarxiv icon

Q-Adapter: Training Your LLM Adapter as a Residual Q-Function

Add code
Jul 04, 2024
Viaarxiv icon

Improving Sample Efficiency of Reinforcement Learning with Background Knowledge from Large Language Models

Add code
Jul 04, 2024
Figure 1 for Improving Sample Efficiency of Reinforcement Learning with Background Knowledge from Large Language Models
Figure 2 for Improving Sample Efficiency of Reinforcement Learning with Background Knowledge from Large Language Models
Figure 3 for Improving Sample Efficiency of Reinforcement Learning with Background Knowledge from Large Language Models
Figure 4 for Improving Sample Efficiency of Reinforcement Learning with Background Knowledge from Large Language Models
Viaarxiv icon

Any-step Dynamics Model Improves Future Predictions for Online and Offline Reinforcement Learning

Add code
May 27, 2024
Viaarxiv icon

BWArea Model: Learning World Model, Inverse Dynamics, and Policy for Controllable Language Generation

Add code
May 27, 2024
Figure 1 for BWArea Model: Learning World Model, Inverse Dynamics, and Policy for Controllable Language Generation
Figure 2 for BWArea Model: Learning World Model, Inverse Dynamics, and Policy for Controllable Language Generation
Figure 3 for BWArea Model: Learning World Model, Inverse Dynamics, and Policy for Controllable Language Generation
Figure 4 for BWArea Model: Learning World Model, Inverse Dynamics, and Policy for Controllable Language Generation
Viaarxiv icon

Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation

Add code
Mar 12, 2024
Figure 1 for Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation
Figure 2 for Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation
Figure 3 for Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation
Figure 4 for Disentangling Policy from Offline Task Representation Learning via Adversarial Data Augmentation
Viaarxiv icon

Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics

Add code
Feb 17, 2024
Figure 1 for Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics
Figure 2 for Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics
Figure 3 for Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics
Figure 4 for Debiased Offline Representation Learning for Fast Online Adaptation in Non-stationary Dynamics
Viaarxiv icon

Imitator Learning: Achieve Out-of-the-Box Imitation Ability in Variable Environments

Add code
Oct 09, 2023
Viaarxiv icon

Policy Regularization with Dataset Constraint for Offline Reinforcement Learning

Add code
Jun 11, 2023
Figure 1 for Policy Regularization with Dataset Constraint for Offline Reinforcement Learning
Figure 2 for Policy Regularization with Dataset Constraint for Offline Reinforcement Learning
Figure 3 for Policy Regularization with Dataset Constraint for Offline Reinforcement Learning
Figure 4 for Policy Regularization with Dataset Constraint for Offline Reinforcement Learning
Viaarxiv icon