Picture for Zhilin Wang

Zhilin Wang

Diverging Preferences: When do Annotators Disagree and do Models Know?

Add code
Oct 18, 2024
Viaarxiv icon

HelpSteer2-Preference: Complementing Ratings with Preferences

Add code
Oct 02, 2024
Viaarxiv icon

Data, Data Everywhere: A Guide for Pretraining Dataset Construction

Add code
Jul 08, 2024
Viaarxiv icon

Nemotron-4 340B Technical Report

Add code
Jun 17, 2024
Viaarxiv icon

HelpSteer2: Open-source dataset for training top-performing reward models

Add code
Jun 12, 2024
Viaarxiv icon

Spotting AI's Touch: Identifying LLM-Paraphrased Spans in Text

Add code
May 21, 2024
Viaarxiv icon

NeMo-Aligner: Scalable Toolkit for Efficient Model Alignment

Add code
May 02, 2024
Viaarxiv icon

HelpSteer: Multi-attribute Helpfulness Dataset for SteerLM

Add code
Nov 16, 2023
Viaarxiv icon

SteerLM: Attribute Conditioned SFT as an (User-Steerable) Alternative to RLHF

Add code
Oct 09, 2023
Viaarxiv icon

Humanoid Agents: Platform for Simulating Human-like Generative Agents

Add code
Oct 09, 2023
Viaarxiv icon