Picture for Janghwan Lee

Janghwan Lee

RILQ: Rank-Insensitive LoRA-based Quantization Error Compensation for Boosting 2-bit Large Language Model Accuracy

Add code
Dec 02, 2024
Viaarxiv icon

AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference

Add code
Nov 15, 2024
Figure 1 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 2 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 3 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 4 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Viaarxiv icon

Efficient Generation of Molecular Clusters with Dual-Scale Equivariant Flow Matching

Add code
Oct 10, 2024
Viaarxiv icon

Improving Conversational Abilities of Quantized Large Language Models via Direct Preference Alignment

Add code
Jul 03, 2024
Viaarxiv icon

SHAPNN: Shapley Value Regularized Tabular Neural Network

Add code
Sep 15, 2023
Viaarxiv icon

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Add code
Aug 13, 2023
Viaarxiv icon

PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices

Add code
May 15, 2023
Viaarxiv icon