Picture for Yuan Hua

Yuan Hua

Improving Generalization of Alignment with Human Preferences through Group Invariant Learning

Add code
Oct 19, 2023
Viaarxiv icon

Secrets of RLHF in Large Language Models Part I: PPO

Add code
Jul 18, 2023
Viaarxiv icon