Picture for Dimitris Papailiopoulos

Dimitris Papailiopoulos

VersaPRM: Multi-Domain Process Reward Model via Synthetic Reasoning Data

Add code
Feb 10, 2025
Viaarxiv icon

Self-Improving Transformers Overcome Easy-to-Hard and Length Generalization Challenges

Add code
Feb 03, 2025
Viaarxiv icon

Task Vectors in In-Context Learning: Emergence, Formation, and Benefit

Add code
Jan 16, 2025
Viaarxiv icon

Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries

Add code
Dec 12, 2024
Figure 1 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Figure 2 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Figure 3 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Figure 4 for Lexico: Extreme KV Cache Compression via Sparse Coding over Universal Dictionaries
Viaarxiv icon

Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition

Add code
Oct 08, 2024
Figure 1 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 2 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 3 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 4 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Viaarxiv icon

From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data

Add code
Jun 27, 2024
Figure 1 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Figure 2 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Figure 3 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Figure 4 for From Artificial Needles to Real Haystacks: Improving Retrieval Capabilities in LLMs by Finetuning on Synthetic Data
Viaarxiv icon

CHAI: Clustered Head Attention for Efficient LLM Inference

Add code
Mar 12, 2024
Figure 1 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 2 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 3 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 4 for CHAI: Clustered Head Attention for Efficient LLM Inference
Viaarxiv icon

How Well Can Transformers Emulate In-context Newton's Method?

Add code
Mar 05, 2024
Viaarxiv icon

Can Mamba Learn How to Learn? A Comparative Study on In-Context Learning Tasks

Add code
Feb 06, 2024
Viaarxiv icon

Looped Transformers are Better at Learning Learning Algorithms

Add code
Nov 21, 2023
Viaarxiv icon