Picture for Zhilin Wang

Zhilin Wang

Lost in Literalism: How Supervised Training Shapes Translationese in LLMs

Add code
Mar 06, 2025
Figure 1 for Lost in Literalism: How Supervised Training Shapes Translationese in LLMs
Figure 2 for Lost in Literalism: How Supervised Training Shapes Translationese in LLMs
Figure 3 for Lost in Literalism: How Supervised Training Shapes Translationese in LLMs
Figure 4 for Lost in Literalism: How Supervised Training Shapes Translationese in LLMs
Viaarxiv icon

Dedicated Feedback and Edit Models Empower Inference-Time Scaling for Open-Ended General-Domain Tasks

Add code
Mar 06, 2025
Viaarxiv icon

Unveiling Attractor Cycles in Large Language Models: A Dynamical Systems View of Successive Paraphrasing

Add code
Feb 21, 2025
Viaarxiv icon

From Drafts to Answers: Unlocking LLM Potential via Aggregation Fine-Tuning

Add code
Jan 21, 2025
Viaarxiv icon

Diverging Preferences: When do Annotators Disagree and do Models Know?

Add code
Oct 18, 2024
Figure 1 for Diverging Preferences: When do Annotators Disagree and do Models Know?
Figure 2 for Diverging Preferences: When do Annotators Disagree and do Models Know?
Figure 3 for Diverging Preferences: When do Annotators Disagree and do Models Know?
Figure 4 for Diverging Preferences: When do Annotators Disagree and do Models Know?
Viaarxiv icon

HelpSteer2-Preference: Complementing Ratings with Preferences

Add code
Oct 02, 2024
Viaarxiv icon

Data, Data Everywhere: A Guide for Pretraining Dataset Construction

Add code
Jul 08, 2024
Viaarxiv icon

Nemotron-4 340B Technical Report

Add code
Jun 17, 2024
Figure 1 for Nemotron-4 340B Technical Report
Figure 2 for Nemotron-4 340B Technical Report
Figure 3 for Nemotron-4 340B Technical Report
Figure 4 for Nemotron-4 340B Technical Report
Viaarxiv icon

HelpSteer2: Open-source dataset for training top-performing reward models

Add code
Jun 12, 2024
Viaarxiv icon

Spotting AI's Touch: Identifying LLM-Paraphrased Spans in Text

Add code
May 21, 2024
Viaarxiv icon