Picture for Yongkweon Jeon

Yongkweon Jeon

Attention-aware Post-training Quantization without Backpropagation

Add code
Jun 19, 2024
Viaarxiv icon

Towards Next-Level Post-Training Quantization of Hyper-Scale Transformers

Add code
Feb 14, 2024
Viaarxiv icon

Genie: Show Me the Data for Quantization

Add code
Dec 09, 2022
Viaarxiv icon

Modulating Regularization Frequency for Efficient Compression-Aware Model Training

Add code
May 05, 2021
Figure 1 for Modulating Regularization Frequency for Efficient Compression-Aware Model Training
Figure 2 for Modulating Regularization Frequency for Efficient Compression-Aware Model Training
Figure 3 for Modulating Regularization Frequency for Efficient Compression-Aware Model Training
Figure 4 for Modulating Regularization Frequency for Efficient Compression-Aware Model Training
Viaarxiv icon

Sequential Encryption of Sparse Neural Networks Toward Optimum Representation of Irregular Sparsity

Add code
May 05, 2021
Figure 1 for Sequential Encryption of Sparse Neural Networks Toward Optimum Representation of Irregular Sparsity
Figure 2 for Sequential Encryption of Sparse Neural Networks Toward Optimum Representation of Irregular Sparsity
Figure 3 for Sequential Encryption of Sparse Neural Networks Toward Optimum Representation of Irregular Sparsity
Figure 4 for Sequential Encryption of Sparse Neural Networks Toward Optimum Representation of Irregular Sparsity
Viaarxiv icon

Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization

Add code
May 05, 2021
Figure 1 for Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Figure 2 for Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Figure 3 for Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Figure 4 for Q-Rater: Non-Convex Optimization for Post-Training Uniform Quantization
Viaarxiv icon

Extremely Low Bit Transformer Quantization for On-Device Neural Machine Translation

Add code
Oct 13, 2020
Figure 1 for Extremely Low Bit Transformer Quantization for On-Device Neural Machine Translation
Figure 2 for Extremely Low Bit Transformer Quantization for On-Device Neural Machine Translation
Figure 3 for Extremely Low Bit Transformer Quantization for On-Device Neural Machine Translation
Figure 4 for Extremely Low Bit Transformer Quantization for On-Device Neural Machine Translation
Viaarxiv icon

FleXOR: Trainable Fractional Quantization

Add code
Sep 09, 2020
Figure 1 for FleXOR: Trainable Fractional Quantization
Figure 2 for FleXOR: Trainable Fractional Quantization
Figure 3 for FleXOR: Trainable Fractional Quantization
Figure 4 for FleXOR: Trainable Fractional Quantization
Viaarxiv icon

BiQGEMM: Matrix Multiplication with Lookup Table For Binary-Coding-based Quantized DNNs

Add code
May 20, 2020
Figure 1 for BiQGEMM: Matrix Multiplication with Lookup Table For Binary-Coding-based Quantized DNNs
Figure 2 for BiQGEMM: Matrix Multiplication with Lookup Table For Binary-Coding-based Quantized DNNs
Figure 3 for BiQGEMM: Matrix Multiplication with Lookup Table For Binary-Coding-based Quantized DNNs
Figure 4 for BiQGEMM: Matrix Multiplication with Lookup Table For Binary-Coding-based Quantized DNNs
Viaarxiv icon