Picture for Hyungjun Kim

Hyungjun Kim

QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference

Add code
Feb 15, 2024
Viaarxiv icon

SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks

Add code
Feb 14, 2024
Viaarxiv icon

Squeezing Large-Scale Diffusion Models for Mobile

Add code
Jul 03, 2023
Viaarxiv icon

OWQ: Lessons learned from activation outliers for weight quantization in large language models

Add code
Jun 13, 2023
Viaarxiv icon

Temporal Dynamic Quantization for Diffusion Models

Add code
Jun 04, 2023
Viaarxiv icon

INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold

Add code
Apr 18, 2022
Figure 1 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 2 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 3 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 4 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Viaarxiv icon

Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution

Add code
Dec 02, 2020
Figure 1 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 2 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 3 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 4 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Viaarxiv icon

Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression

Add code
Sep 08, 2020
Figure 1 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 2 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 3 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 4 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Viaarxiv icon

BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations

Add code
Feb 16, 2020
Figure 1 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 2 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 3 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 4 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Viaarxiv icon

Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays

Add code
Aug 02, 2019
Figure 1 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Figure 2 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Figure 3 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Figure 4 for Zero-shifting Technique for Deep Neural Network Training on Resistive Cross-point Arrays
Viaarxiv icon