Picture for Hanze Dong

Hanze Dong

BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation

Add code
Feb 06, 2025
Figure 1 for BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
Figure 2 for BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
Figure 3 for BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
Figure 4 for BOLT: Bootstrap Long Chain-of-Thought in Language Models without Distillation
Viaarxiv icon

Reward-Guided Speculative Decoding for Efficient LLM Reasoning

Add code
Jan 31, 2025
Figure 1 for Reward-Guided Speculative Decoding for Efficient LLM Reasoning
Figure 2 for Reward-Guided Speculative Decoding for Efficient LLM Reasoning
Figure 3 for Reward-Guided Speculative Decoding for Efficient LLM Reasoning
Figure 4 for Reward-Guided Speculative Decoding for Efficient LLM Reasoning
Viaarxiv icon

Offline Reinforcement Learning for LLM Multi-Step Reasoning

Add code
Dec 20, 2024
Viaarxiv icon

Entropy-Regularized Process Reward Model

Add code
Dec 15, 2024
Viaarxiv icon

Automatic Curriculum Expert Iteration for Reliable LLM Reasoning

Add code
Oct 10, 2024
Figure 1 for Automatic Curriculum Expert Iteration for Reliable LLM Reasoning
Figure 2 for Automatic Curriculum Expert Iteration for Reliable LLM Reasoning
Figure 3 for Automatic Curriculum Expert Iteration for Reliable LLM Reasoning
Figure 4 for Automatic Curriculum Expert Iteration for Reliable LLM Reasoning
Viaarxiv icon

MathHay: An Automated Benchmark for Long-Context Mathematical Reasoning in LLMs

Add code
Oct 07, 2024
Figure 1 for MathHay: An Automated Benchmark for Long-Context Mathematical Reasoning in LLMs
Figure 2 for MathHay: An Automated Benchmark for Long-Context Mathematical Reasoning in LLMs
Figure 3 for MathHay: An Automated Benchmark for Long-Context Mathematical Reasoning in LLMs
Figure 4 for MathHay: An Automated Benchmark for Long-Context Mathematical Reasoning in LLMs
Viaarxiv icon

FIRST: Teach A Reliable Large Language Model Through Efficient Trustworthy Distillation

Add code
Aug 22, 2024
Viaarxiv icon

ThinK: Thinner Key Cache by Query-Driven Pruning

Add code
Jul 30, 2024
Viaarxiv icon

Faster Sampling via Stochastic Gradient Proximal Sampler

Add code
May 27, 2024
Viaarxiv icon

Reverse Transition Kernel: A Flexible Framework to Accelerate Diffusion Inference

Add code
May 26, 2024
Figure 1 for Reverse Transition Kernel: A Flexible Framework to Accelerate Diffusion Inference
Figure 2 for Reverse Transition Kernel: A Flexible Framework to Accelerate Diffusion Inference
Figure 3 for Reverse Transition Kernel: A Flexible Framework to Accelerate Diffusion Inference
Viaarxiv icon