Picture for Dimitris Papailiopoulos

Dimitris Papailiopoulos

Endless Terminals: Scaling RL Environments for Terminal Agents

Add code
Jan 27, 2026
Viaarxiv icon

Wait, Wait, Wait... Why Do Reasoning Models Loop?

Add code
Dec 15, 2025
Viaarxiv icon

Extrapolation by Association: Length Generalization Transfer in Transformers

Add code
Jun 10, 2025
Viaarxiv icon

Phi-4-reasoning Technical Report

Add code
Apr 30, 2025
Viaarxiv icon

VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data

Add code
Feb 10, 2025
Figure 1 for VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data
Figure 2 for VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data
Figure 3 for VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data
Figure 4 for VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data
Viaarxiv icon

Self-Improving Transformers Overcome Easy-to-Hard and Length Generalization Challenges

Add code
Feb 03, 2025
Viaarxiv icon

Task Vectors in In-Context Learning: Emergence, Formation, and Benefit

Add code
Jan 16, 2025
Viaarxiv icon

Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries

Add code
Dec 12, 2024
Figure 1 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Figure 2 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Figure 3 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Figure 4 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Viaarxiv icon

Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition

Add code
Oct 08, 2024
Figure 1 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 2 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 3 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 4 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Viaarxiv icon

From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data

Add code
Jun 27, 2024
Figure 1 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Figure 2 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Figure 3 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Figure 4 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Viaarxiv icon