Picture for Congliang Chen

Congliang Chen

Off-Policy Value-Based Reinforcement Learning for Large Language Models

Add code
Mar 24, 2026
Viaarxiv icon

Adam Converges Without Any Modification On Update Rules

Add code
Mar 02, 2026
Viaarxiv icon

ORGEval: Graph-Theoretic Evaluation of LLMs in Optimization Modeling

Add code
Oct 31, 2025
Viaarxiv icon

Rethinking Data Mixture for Large Language Models: A Comprehensive Survey and New Perspectives

Add code
May 27, 2025
Viaarxiv icon

Exploring the Generalization Capabilities of AID-based Bi-level Optimization

Add code
Nov 25, 2024
Viaarxiv icon

Entropic Distribution Matching in Supervised Fine-tuning of LLMs: Less Overfitting and Better Diversity

Add code
Aug 29, 2024
Viaarxiv icon

Adam-mini: Use Fewer Learning Rates To Gain More

Add code
Jun 26, 2024
Viaarxiv icon

Why Transformers Need Adam: A Hessian Perspective

Add code
Feb 26, 2024
Figure 1 for Why Transformers Need Adam: A Hessian Perspective
Figure 2 for Why Transformers Need Adam: A Hessian Perspective
Figure 3 for Why Transformers Need Adam: A Hessian Perspective
Figure 4 for Why Transformers Need Adam: A Hessian Perspective
Viaarxiv icon

Rethinking SIGN Training: Provable Nonconvex Acceleration without First- and Second-Order Gradient Lipschitz

Add code
Oct 23, 2023
Figure 1 for Rethinking SIGN Training: Provable Nonconvex Acceleration without First- and Second-Order Gradient Lipschitz
Figure 2 for Rethinking SIGN Training: Provable Nonconvex Acceleration without First- and Second-Order Gradient Lipschitz
Figure 3 for Rethinking SIGN Training: Provable Nonconvex Acceleration without First- and Second-Order Gradient Lipschitz
Figure 4 for Rethinking SIGN Training: Provable Nonconvex Acceleration without First- and Second-Order Gradient Lipschitz
Viaarxiv icon

Adam Can Converge Without Any Modification on Update Rules

Add code
Aug 23, 2022
Figure 1 for Adam Can Converge Without Any Modification on Update Rules
Figure 2 for Adam Can Converge Without Any Modification on Update Rules
Figure 3 for Adam Can Converge Without Any Modification on Update Rules
Figure 4 for Adam Can Converge Without Any Modification on Update Rules
Viaarxiv icon