Picture for Tian Liang

Tian Liang

Do NOT Think That Much for 2+3=? On the Overthinking of o1-Like LLMs

Add code
Dec 30, 2024
Viaarxiv icon

Teaching LLMs to Refine with Tools

Add code
Dec 22, 2024
Viaarxiv icon

Critical Tokens Matter: Token-Level Contrastive Estimation Enhances LLM's Reasoning Capability

Add code
Dec 02, 2024
Viaarxiv icon

Draft Model Knows When to Stop: A Self-Verification Length Policy for Speculative Decoding

Add code
Nov 27, 2024
Viaarxiv icon

Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training

Add code
Jul 12, 2024
Figure 1 for Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training
Figure 2 for Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training
Figure 3 for Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training
Figure 4 for Refuse Whenever You Feel Unsafe: Improving Safety in LLMs via Decoupled Refusal Training
Viaarxiv icon

How Far Are We on the Decision-Making of LLMs? Evaluating LLMs' Gaming Ability in Multi-Agent Environments

Add code
Mar 18, 2024
Viaarxiv icon

CriticBench: Benchmarking LLMs for Critique-Correct Reasoning

Add code
Mar 08, 2024
Viaarxiv icon

Leveraging Word Guessing Games to Assess the Intelligence of Large Language Models

Add code
Nov 06, 2023
Viaarxiv icon

Encouraging Divergent Thinking in Large Language Models through Multi-Agent Debate

Add code
May 30, 2023
Viaarxiv icon

Exploring Human-Like Translation Strategy with Large Language Models

Add code
May 06, 2023
Viaarxiv icon