Picture for Andrei Panferov

Andrei Panferov

Quartet II: Accurate LLM Pre-Training in NVFP4 by Improved Unbiased Gradient Estimation

Add code
Jan 30, 2026
Viaarxiv icon

Apertus: Democratizing Open and Compliant LLMs for Global Language Environments

Add code
Sep 17, 2025
Figure 1 for Apertus: Democratizing Open and Compliant LLMs for Global Language Environments
Figure 2 for Apertus: Democratizing Open and Compliant LLMs for Global Language Environments
Figure 3 for Apertus: Democratizing Open and Compliant LLMs for Global Language Environments
Figure 4 for Apertus: Democratizing Open and Compliant LLMs for Global Language Environments
Viaarxiv icon

Quartet: Native FP4 Training Can Be Optimal for Large Language Models

Add code
May 20, 2025
Figure 1 for Quartet: Native FP4 Training Can Be Optimal for Large Language Models
Figure 2 for Quartet: Native FP4 Training Can Be Optimal for Large Language Models
Figure 3 for Quartet: Native FP4 Training Can Be Optimal for Large Language Models
Figure 4 for Quartet: Native FP4 Training Can Be Optimal for Large Language Models
Viaarxiv icon

QuEST: Stable Training of LLMs with 1-Bit Weights and Activations

Add code
Feb 07, 2025
Figure 1 for QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
Figure 2 for QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
Figure 3 for QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
Figure 4 for QuEST: Stable Training of LLMs with 1-Bit Weights and Activations
Viaarxiv icon

Pushing the Limits of Large Language Model Quantization via the Linearity Theorem

Add code
Nov 26, 2024
Figure 1 for Pushing the Limits of Large Language Model Quantization via the Linearity Theorem
Figure 2 for Pushing the Limits of Large Language Model Quantization via the Linearity Theorem
Figure 3 for Pushing the Limits of Large Language Model Quantization via the Linearity Theorem
Figure 4 for Pushing the Limits of Large Language Model Quantization via the Linearity Theorem
Viaarxiv icon

Extreme Compression of Large Language Models via Additive Quantization

Add code
Jan 11, 2024
Figure 1 for Extreme Compression of Large Language Models via Additive Quantization
Figure 2 for Extreme Compression of Large Language Models via Additive Quantization
Figure 3 for Extreme Compression of Large Language Models via Additive Quantization
Figure 4 for Extreme Compression of Large Language Models via Additive Quantization
Viaarxiv icon

Correlated Quantization for Faster Nonconvex Distributed Optimization

Add code
Jan 10, 2024
Viaarxiv icon