Picture for Sukjin Hong

Sukjin Hong

RILQ: Rank-Insensitive LoRA-based Quantization Error Compensation for Boosting 2-bit Large Language Model Accuracy

Add code
Dec 02, 2024
Viaarxiv icon

Improving Conversational Abilities of Quantized Large Language Models via Direct Preference Alignment

Add code
Jul 03, 2024
Viaarxiv icon

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Add code
Aug 13, 2023
Viaarxiv icon

Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective

Add code
Feb 03, 2023
Viaarxiv icon

Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders

Add code
Nov 20, 2022
Viaarxiv icon