Picture for Haocheng Xi

Haocheng Xi

Oscillation-Reduced MXFP4 Training for Vision Transformers

Add code
Feb 28, 2025
Viaarxiv icon

SpargeAttn: Accurate Sparse Attention Accelerating Any Model Inference

Add code
Feb 25, 2025
Viaarxiv icon

Sparse VideoGen: Accelerating Video Diffusion Transformers with Spatial-Temporal Sparsity

Add code
Feb 03, 2025
Figure 1 for Sparse VideoGen: Accelerating Video Diffusion Transformers with Spatial-Temporal Sparsity
Figure 2 for Sparse VideoGen: Accelerating Video Diffusion Transformers with Spatial-Temporal Sparsity
Figure 3 for Sparse VideoGen: Accelerating Video Diffusion Transformers with Spatial-Temporal Sparsity
Figure 4 for Sparse VideoGen: Accelerating Video Diffusion Transformers with Spatial-Temporal Sparsity
Viaarxiv icon

NVILA: Efficient Frontier Visual Language Models

Add code
Dec 05, 2024
Figure 1 for NVILA: Efficient Frontier Visual Language Models
Figure 2 for NVILA: Efficient Frontier Visual Language Models
Figure 3 for NVILA: Efficient Frontier Visual Language Models
Figure 4 for NVILA: Efficient Frontier Visual Language Models
Viaarxiv icon

COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Training

Add code
Oct 25, 2024
Figure 1 for COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Training
Figure 2 for COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Training
Figure 3 for COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Training
Figure 4 for COAT: Compressing Optimizer states and Activation for Memory-Efficient FP8 Training
Viaarxiv icon

Jetfire: Efficient and Accurate Transformer Pretraining with INT8 Data Flow and Per-Block Quantization

Add code
Mar 19, 2024
Viaarxiv icon

T-Rex: Text-assisted Retrosynthesis Prediction

Add code
Jan 26, 2024
Viaarxiv icon

Training Transformers with 4-bit Integers

Add code
Jun 22, 2023
Viaarxiv icon