Picture for Hyungjun Kim

Hyungjun Kim

Debunking the CUDA Myth Towards GPU-based AI Systems

Add code
Dec 31, 2024
Figure 1 for Debunking the CUDA Myth Towards GPU-based AI Systems
Figure 2 for Debunking the CUDA Myth Towards GPU-based AI Systems
Figure 3 for Debunking the CUDA Myth Towards GPU-based AI Systems
Figure 4 for Debunking the CUDA Myth Towards GPU-based AI Systems
Viaarxiv icon

QUICK: Quantization-aware Interleaving and Conflict-free Kernel for efficient LLM inference

Add code
Feb 15, 2024
Viaarxiv icon

SLEB: Streamlining LLMs through Redundancy Verification and Elimination of Transformer Blocks

Add code
Feb 14, 2024
Viaarxiv icon

Squeezing Large-Scale Diffusion Models for Mobile

Add code
Jul 03, 2023
Viaarxiv icon

OWQ: Lessons learned from activation outliers for weight quantization in large language models

Add code
Jun 13, 2023
Figure 1 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Figure 2 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Figure 3 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Figure 4 for OWQ: Lessons learned from activation outliers for weight quantization in large language models
Viaarxiv icon

Temporal Dynamic Quantization for Diffusion Models

Add code
Jun 04, 2023
Viaarxiv icon

INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold

Add code
Apr 18, 2022
Figure 1 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 2 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 3 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Figure 4 for INSTA-BNN: Binary Neural Network with INSTAnce-aware Threshold
Viaarxiv icon

Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution

Add code
Dec 02, 2020
Figure 1 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 2 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 3 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Figure 4 for Improving Accuracy of Binary Neural Networks using Unbalanced Activation Distribution
Viaarxiv icon

Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression

Add code
Sep 08, 2020
Figure 1 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 2 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 3 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Figure 4 for Empirical Strategy for Stretching Probability Distribution in Neural-network-based Regression
Viaarxiv icon

BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations

Add code
Feb 16, 2020
Figure 1 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 2 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 3 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Figure 4 for BinaryDuo: Reducing Gradient Mismatch in Binary Activation Network by Coupling Binary Activations
Viaarxiv icon