Picture for Seungwoo Son

Seungwoo Son

Prefixing Attention Sinks can Mitigate Activation Outliers for Large Language Model Quantization

Add code
Jun 17, 2024
Viaarxiv icon

MaskedKD: Efficient Distillation of Vision Transformers with Masked Images

Add code
Feb 21, 2023
Viaarxiv icon