Picture for Wenyu Du

Wenyu Du

The Station: An Open-World Environment for AI-Driven Discovery

Add code
Nov 09, 2025
Figure 1 for The Station: An Open-World Environment for AI-Driven Discovery
Figure 2 for The Station: An Open-World Environment for AI-Driven Discovery
Figure 3 for The Station: An Open-World Environment for AI-Driven Discovery
Figure 4 for The Station: An Open-World Environment for AI-Driven Discovery
Viaarxiv icon

Thinker: Learning to Think Fast and Slow

Add code
May 27, 2025
Figure 1 for Thinker: Learning to Think Fast and Slow
Figure 2 for Thinker: Learning to Think Fast and Slow
Figure 3 for Thinker: Learning to Think Fast and Slow
Figure 4 for Thinker: Learning to Think Fast and Slow
Viaarxiv icon

Learning from Peers in Reasoning Models

Add code
May 12, 2025
Viaarxiv icon

Learning from Failures in Multi-Attempt Reinforcement Learning

Add code
Mar 04, 2025
Viaarxiv icon

Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking

Add code
Feb 27, 2025
Figure 1 for Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking
Figure 2 for Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking
Figure 3 for Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking
Figure 4 for Finite State Automata Inside Transformers with Chain-of-Thought: A Mechanistic Study on State Tracking
Viaarxiv icon

Towards Understanding Fine-Tuning Mechanisms of LLMs via Circuit Analysis

Add code
Feb 17, 2025
Viaarxiv icon

Unlocking Continual Learning Abilities in Language Models

Add code
Jun 25, 2024
Figure 1 for Unlocking Continual Learning Abilities in Language Models
Figure 2 for Unlocking Continual Learning Abilities in Language Models
Figure 3 for Unlocking Continual Learning Abilities in Language Models
Figure 4 for Unlocking Continual Learning Abilities in Language Models
Viaarxiv icon

Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training

Add code
May 24, 2024
Figure 1 for Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
Figure 2 for Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
Figure 3 for Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
Figure 4 for Stacking Your Transformers: A Closer Look at Model Growth for Efficient LLM Pre-Training
Viaarxiv icon

m2mKD: Module-to-Module Knowledge Distillation for Modular Transformers

Add code
Feb 26, 2024
Figure 1 for m2mKD: Module-to-Module Knowledge Distillation for Modular Transformers
Figure 2 for m2mKD: Module-to-Module Knowledge Distillation for Modular Transformers
Figure 3 for m2mKD: Module-to-Module Knowledge Distillation for Modular Transformers
Figure 4 for m2mKD: Module-to-Module Knowledge Distillation for Modular Transformers
Viaarxiv icon

f-Divergence Minimization for Sequence-Level Knowledge Distillation

Add code
Jul 27, 2023
Figure 1 for f-Divergence Minimization for Sequence-Level Knowledge Distillation
Figure 2 for f-Divergence Minimization for Sequence-Level Knowledge Distillation
Figure 3 for f-Divergence Minimization for Sequence-Level Knowledge Distillation
Figure 4 for f-Divergence Minimization for Sequence-Level Knowledge Distillation
Viaarxiv icon