Picture for Yaodong Yang

Yaodong Yang

Random Feature Models with Learnable Activation Functions

Add code
Nov 29, 2024
Viaarxiv icon

Object-Centric Dexterous Manipulation from Human Motion Data

Add code
Nov 06, 2024
Figure 1 for Object-Centric Dexterous Manipulation from Human Motion Data
Figure 2 for Object-Centric Dexterous Manipulation from Human Motion Data
Figure 3 for Object-Centric Dexterous Manipulation from Human Motion Data
Figure 4 for Object-Centric Dexterous Manipulation from Human Motion Data
Viaarxiv icon

Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping

Add code
Oct 30, 2024
Figure 1 for Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping
Figure 2 for Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping
Figure 3 for Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping
Figure 4 for Neural Attention Field: Emerging Point Relevance in 3D Scenes for One-Shot Dexterous Grasping
Viaarxiv icon

Magnetic Preference Optimization: Achieving Last-iterate Convergence for Language Models Alignment

Add code
Oct 22, 2024
Viaarxiv icon

Computing Ex Ante Equilibrium in Heterogeneous Zero-Sum Team Games

Add code
Oct 02, 2024
Viaarxiv icon

Sequence to Sequence Reward Modeling: Improving RLHF by Language Feedback

Add code
Aug 30, 2024
Viaarxiv icon

A Survey on Self-play Methods in Reinforcement Learning

Add code
Aug 02, 2024
Viaarxiv icon

ProgressGym: Alignment with a Millennium of Moral Progress

Add code
Jun 28, 2024
Figure 1 for ProgressGym: Alignment with a Millennium of Moral Progress
Figure 2 for ProgressGym: Alignment with a Millennium of Moral Progress
Figure 3 for ProgressGym: Alignment with a Millennium of Moral Progress
Figure 4 for ProgressGym: Alignment with a Millennium of Moral Progress
Viaarxiv icon

PKU-SafeRLHF: A Safety Alignment Preference Dataset for Llama Family Models

Add code
Jun 20, 2024
Figure 1 for PKU-SafeRLHF: A Safety Alignment Preference Dataset for Llama Family Models
Figure 2 for PKU-SafeRLHF: A Safety Alignment Preference Dataset for Llama Family Models
Figure 3 for PKU-SafeRLHF: A Safety Alignment Preference Dataset for Llama Family Models
Figure 4 for PKU-SafeRLHF: A Safety Alignment Preference Dataset for Llama Family Models
Viaarxiv icon

SafeSora: Towards Safety Alignment of Text2Video Generation via a Human Preference Dataset

Add code
Jun 20, 2024
Viaarxiv icon