Picture for Ron Banner

Ron Banner

EXAQ: Exponent Aware Quantization For LLMs Acceleration

Add code
Oct 04, 2024
Figure 1 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 2 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 3 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Figure 4 for EXAQ: Exponent Aware Quantization For LLMs Acceleration
Viaarxiv icon

DropCompute: simple and more robust distributed synchronous training via compute variance reduction

Add code
Jun 18, 2023
Viaarxiv icon

Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients

Add code
Mar 21, 2022
Figure 1 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 2 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 3 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Figure 4 for Optimal Fine-Grained N:M sparsity for Activations and Neural Gradients
Viaarxiv icon

Energy awareness in low precision neural networks

Add code
Feb 06, 2022
Viaarxiv icon

On Recoverability of Graph Neural Network Representations

Add code
Jan 30, 2022
Figure 1 for On Recoverability of Graph Neural Network Representations
Figure 2 for On Recoverability of Graph Neural Network Representations
Figure 3 for On Recoverability of Graph Neural Network Representations
Figure 4 for On Recoverability of Graph Neural Network Representations
Viaarxiv icon

Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning

Add code
Dec 19, 2021
Figure 1 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 2 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 3 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Figure 4 for Logarithmic Unbiased Quantization: Practical 4-bit Training in Deep Learning
Viaarxiv icon

Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks

Add code
Feb 16, 2021
Figure 1 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Figure 2 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Figure 3 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Figure 4 for Accelerated Sparse Neural Training: A Provable and Efficient Method to Find N:M Transposable Masks
Viaarxiv icon

GAN Steerability without optimization

Add code
Dec 09, 2020
Figure 1 for GAN Steerability without optimization
Figure 2 for GAN Steerability without optimization
Figure 3 for GAN Steerability without optimization
Figure 4 for GAN Steerability without optimization
Viaarxiv icon

Neural gradients are lognormally distributed: understanding sparse and quantized training

Add code
Jun 17, 2020
Figure 1 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Figure 2 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Figure 3 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Figure 4 for Neural gradients are lognormally distributed: understanding sparse and quantized training
Viaarxiv icon

Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming

Add code
Jun 14, 2020
Figure 1 for Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming
Figure 2 for Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming
Figure 3 for Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming
Figure 4 for Improving Post Training Neural Quantization: Layer-wise Calibration and Integer Programming
Viaarxiv icon