Picture for Sukjin Hong

Sukjin Hong

Improving Conversational Abilities of Quantized Large Language Models via Direct Preference Alignment

Add code
Jul 03, 2024
Viaarxiv icon

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Add code
Aug 13, 2023
Viaarxiv icon

Revisiting Intermediate Layer Distillation for Compressing Language Models: An Overfitting Perspective

Add code
Feb 03, 2023
Viaarxiv icon

Understanding and Improving Knowledge Distillation for Quantization-Aware Training of Large Transformer Encoders

Add code
Nov 20, 2022
Viaarxiv icon