Picture for Jungwook Choi

Jungwook Choi

Quantization-Aware Imitation-Learning for Resource-Efficient Robotic Control

Add code
Dec 02, 2024
Viaarxiv icon

RILQ: Rank-Insensitive LoRA-based Quantization Error Compensation for Boosting 2-bit Large Language Model Accuracy

Add code
Dec 02, 2024
Viaarxiv icon

AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference

Add code
Nov 15, 2024
Figure 1 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 2 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 3 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Figure 4 for AMXFP4: Taming Activation Outliers with Asymmetric Microscaling Floating-Point for 4-bit LLM Inference
Viaarxiv icon

InfiniPot: Infinite Context Processing on Memory-Constrained LLMs

Add code
Oct 02, 2024
Viaarxiv icon

Selectively Dilated Convolution for Accuracy-Preserving Sparse Pillar-based Embedded 3D Object Detection

Add code
Aug 25, 2024
Viaarxiv icon

Improving Conversational Abilities of Quantized Large Language Models via Direct Preference Alignment

Add code
Jul 03, 2024
Viaarxiv icon

Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization

Add code
Nov 09, 2023
Viaarxiv icon

Token-Scaled Logit Distillation for Ternary Weight Generative Language Models

Add code
Aug 13, 2023
Viaarxiv icon

PillarAcc: Sparse PointPillars Accelerator for Real-Time Point Cloud 3D Object Detection on Edge Devices

Add code
May 15, 2023
Viaarxiv icon

Teacher Intervention: Improving Convergence of Quantization Aware Training for Ultra-Low Precision Transformers

Add code
Feb 23, 2023
Viaarxiv icon