Picture for Shuang Qiu

Shuang Qiu

University of Michigan, Ann Arbor

Forward KL Regularized Preference Optimization for Aligning Diffusion Policies

Add code
Sep 09, 2024
Figure 1 for Forward KL Regularized Preference Optimization for Aligning Diffusion Policies
Figure 2 for Forward KL Regularized Preference Optimization for Aligning Diffusion Policies
Figure 3 for Forward KL Regularized Preference Optimization for Aligning Diffusion Policies
Figure 4 for Forward KL Regularized Preference Optimization for Aligning Diffusion Policies
Viaarxiv icon

Traversing Pareto Optimal Policies: Provably Efficient Multi-Objective Reinforcement Learning

Add code
Jul 24, 2024
Viaarxiv icon

Pessimism Meets Risk: Risk-Sensitive Offline Reinforcement Learning

Add code
Jul 10, 2024
Viaarxiv icon

Human-like object concept representations emerge naturally in multimodal large language models

Add code
Jul 01, 2024
Viaarxiv icon

Arithmetic Control of LLMs for Diverse User Preferences: Directional Preference Alignment with Multi-Objective Rewards

Add code
Mar 06, 2024
Viaarxiv icon

Rewards-in-Context: Multi-objective Alignment of Foundation Models with Dynamic Preference Adjustment

Add code
Feb 25, 2024
Viaarxiv icon

A Temporal-Spectral Fusion Transformer with Subject-specific Adapter for Enhancing RSVP-BCI Decoding

Add code
Jan 12, 2024
Viaarxiv icon

Posterior Sampling for Competitive RL: Function Approximation and Partial Observation

Add code
Oct 30, 2023
Viaarxiv icon

StairNetV3: Depth-aware Stair Modeling using Deep Learning

Add code
Aug 13, 2023
Viaarxiv icon

On the Value of Myopic Behavior in Policy Reuse

Add code
May 28, 2023
Viaarxiv icon