Picture for Dharmendra S. Modha

Dharmendra S. Modha

Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference

Add code
Jan 30, 2023
Figure 1 for Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Figure 2 for Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Figure 3 for Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Figure 4 for Efficient and Effective Methods for Mixed Precision Neural Network Quantization for Faster, Energy-efficient Inference
Viaarxiv icon

Learned Step Size Quantization

Add code
Feb 21, 2019
Figure 1 for Learned Step Size Quantization
Figure 2 for Learned Step Size Quantization
Figure 3 for Learned Step Size Quantization
Figure 4 for Learned Step Size Quantization
Viaarxiv icon

Low Precision Policy Distillation with Application to Low-Power, Real-time Sensation-Cognition-Action Loop with Neuromorphic Computing

Add code
Sep 25, 2018
Figure 1 for Low Precision Policy Distillation with Application to Low-Power, Real-time Sensation-Cognition-Action Loop with Neuromorphic Computing
Figure 2 for Low Precision Policy Distillation with Application to Low-Power, Real-time Sensation-Cognition-Action Loop with Neuromorphic Computing
Figure 3 for Low Precision Policy Distillation with Application to Low-Power, Real-time Sensation-Cognition-Action Loop with Neuromorphic Computing
Figure 4 for Low Precision Policy Distillation with Application to Low-Power, Real-time Sensation-Cognition-Action Loop with Neuromorphic Computing
Viaarxiv icon

Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference

Add code
Sep 11, 2018
Figure 1 for Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference
Figure 2 for Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference
Figure 3 for Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference
Figure 4 for Discovering Low-Precision Networks Close to Full-Precision Networks for Efficient Embedded Inference
Viaarxiv icon

Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing

Add code
May 24, 2016
Figure 1 for Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
Figure 2 for Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
Figure 3 for Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
Figure 4 for Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
Viaarxiv icon

Mapping Generative Models onto a Network of Digital Spiking Neurons

Add code
Oct 09, 2015
Figure 1 for Mapping Generative Models onto a Network of Digital Spiking Neurons
Figure 2 for Mapping Generative Models onto a Network of Digital Spiking Neurons
Figure 3 for Mapping Generative Models onto a Network of Digital Spiking Neurons
Figure 4 for Mapping Generative Models onto a Network of Digital Spiking Neurons
Viaarxiv icon