Picture for Brian Chmiel

Brian Chmiel

EXAQ: Exponent Aware Quantization For LLMs Acceleration

Add code
Oct 04, 2024
Figure 1 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 2 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 3 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 4 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Viaarxiv icon

Bimodal Distributed Binarized Neural Networks

Add code
Apr 05, 2022
Figure 1 for Bimodal Distributed Binarized Neural Networks
Figure 2 for Bimodal Distributed Binarized Neural Networks
Figure 3 for Bimodal Distributed Binarized Neural Networks
Figure 4 for Bimodal Distributed Binarized Neural Networks
Viaarxiv icon

Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients

Add code
Mar 21, 2022
Figure 1 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 2 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 3 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 4 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Viaarxiv icon

Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning

Add code
Dec 19, 2021
Figure 1 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 2 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 3 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 4 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Viaarxiv icon

Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks

Add code
Feb 16, 2021
Figure 1 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Figure 2 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Figure 3 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Figure 4 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Viaarxiv icon

Neural gradients are lognormally distributed: understanding sparse and quantized training

Add code
Jun 17, 2020
Figure 1 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Figure 2 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Figure 3 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Figure 4 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Viaarxiv icon

Colored Noise Injection for Training Adversarially Robust Neural Networks

Add code
Mar 20, 2020
Figure 1 for Colored Noise Injection for Training Adversarially Robust Neural Networks
Figure 2 for Colored Noise Injection for Training Adversarially Robust Neural Networks
Viaarxiv icon

Robust Quantization: One Model to Rule Them All

Add code
Feb 18, 2020
Figure 1 for Robust Quantization: One Model to Rule Them All
Figure 2 for Robust Quantization: One Model to Rule Them All
Figure 3 for Robust Quantization: One Model to Rule Them All
Figure 4 for Robust Quantization: One Model to Rule Them All
Viaarxiv icon

Smoothed Inference for Adversarially-Trained Models

Add code
Nov 17, 2019
Figure 1 for Smoothed Inference for Adversarially-Trained Models
Figure 2 for Smoothed Inference for Adversarially-Trained Models
Figure 3 for Smoothed Inference for Adversarially-Trained Models
Figure 4 for Smoothed Inference for Adversarially-Trained Models
Viaarxiv icon

Loss Aware Post-training Quantization

Add code
Nov 17, 2019
Figure 1 for Loss Aware Post-training Quantization
Figure 2 for Loss Aware Post-training Quantization
Figure 3 for Loss Aware Post-training Quantization
Figure 4 for Loss Aware Post-training Quantization
Viaarxiv icon