Picture for Seungwoo Son

Seungwoo Son

TurboBoA: Faster and Exact Attention-aware Quantization without Backpropagation

Add code
Feb 04, 2026
Viaarxiv icon

Two-Stage Grid Optimization for Group-wise Quantization of LLMs

Add code
Feb 02, 2026
Viaarxiv icon

Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization

Add code
Jun 17, 2024
Figure 1 for Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization
Figure 2 for Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization
Figure 3 for Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization
Figure 4 for Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization
Viaarxiv icon

MaskedKD: Efficient Distillation of Vision Transformers with Masked Images

Add code
Feb 21, 2023
Viaarxiv icon