Picture for Dan Alistarh

Dan Alistarh

MatGPTQ: Accurate and Efficient Post-Training Matryoshka Quantization

Add code
Feb 03, 2026
Viaarxiv icon

DASH: Faster Shampoo via Batched Block Preconditioning and Efficient Inverse-Root Solvers

Add code
Feb 02, 2026
Viaarxiv icon

Behemoth: Benchmarking Unlearning in LLMs Using Fully Synthetic Data

Add code
Jan 30, 2026
Viaarxiv icon

Quartet II: Accurate LLM Pre-Training in NVFP4 by Improved Unbiased Gradient Estimation

Add code
Jan 30, 2026
Viaarxiv icon

ECO: Quantized Training without Full-Precision Master Weights

Add code
Jan 29, 2026
Viaarxiv icon

LLMQ: Efficient Lower-Precision Pretraining for Consumer GPUs

Add code
Dec 17, 2025
Viaarxiv icon

Speculative Decoding Speed-of-Light: Optimal Lower Bounds via Branching Random Walks

Add code
Dec 12, 2025
Viaarxiv icon

Expand Neurons, Not Parameters

Add code
Oct 06, 2025
Figure 1 for Expand Neurons, Not Parameters
Figure 2 for Expand Neurons, Not Parameters
Figure 3 for Expand Neurons, Not Parameters
Figure 4 for Expand Neurons, Not Parameters
Viaarxiv icon

The Unseen Frontier: Pushing the Limits of LLM Sparsity with Surrogate-Free ADMM

Add code
Oct 02, 2025
Viaarxiv icon

Optimizers Qualitatively Alter Solutions And We Should Leverage This

Add code
Jul 16, 2025
Viaarxiv icon