Picture for Tianyu Pang

Tianyu Pang

Rethinking the Trust Region in LLM Reinforcement Learning

Add code
Feb 04, 2026
Viaarxiv icon

Depth, Not Data: An Analysis of Hessian Spectral Bifurcation

Add code
Jan 31, 2026
Viaarxiv icon

Suspicious Alignment of SGD: A Fine-Grained Step Size Condition Analysis

Add code
Jan 16, 2026
Viaarxiv icon

Demystifying the Slash Pattern in Attention: The Role of RoPE

Add code
Jan 13, 2026
Viaarxiv icon

Defeating the Training-Inference Mismatch via FP16

Add code
Oct 30, 2025
Viaarxiv icon

Nonparametric Data Attribution for Diffusion Models

Add code
Oct 16, 2025
Viaarxiv icon

Imperceptible Jailbreaking against Large Language Models

Add code
Oct 06, 2025
Viaarxiv icon

Variational Reasoning for Language Models

Add code
Sep 26, 2025
Figure 1 for Variational Reasoning for Language Models
Figure 2 for Variational Reasoning for Language Models
Figure 3 for Variational Reasoning for Language Models
Figure 4 for Variational Reasoning for Language Models
Viaarxiv icon

Language Models Can Learn from Verbal Feedback Without Scalar Rewards

Add code
Sep 26, 2025
Figure 1 for Language Models Can Learn from Verbal Feedback Without Scalar Rewards
Figure 2 for Language Models Can Learn from Verbal Feedback Without Scalar Rewards
Figure 3 for Language Models Can Learn from Verbal Feedback Without Scalar Rewards
Figure 4 for Language Models Can Learn from Verbal Feedback Without Scalar Rewards
Viaarxiv icon

Why LLM Safety Guardrails Collapse After Fine-tuning: A Similarity Analysis Between Alignment and Fine-tuning Datasets

Add code
Jun 05, 2025
Viaarxiv icon