Picture for Ruizhou Ding

Ruizhou Ding

Shawn

QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration

Add code
Jun 30, 2022
Figure 1 for QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration
Figure 2 for QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration
Figure 3 for QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration
Figure 4 for QUIDAM: A Framework for Quantization-Aware DNN Accelerator and Model Co-Exploration
Viaarxiv icon

QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality

Add code
May 20, 2022
Figure 1 for QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality
Figure 2 for QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality
Figure 3 for QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality
Figure 4 for QADAM: Quantization-Aware DNN Accelerator Modeling for Pareto-Optimality
Viaarxiv icon

QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators

Add code
May 17, 2022
Figure 1 for QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators
Figure 2 for QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators
Figure 3 for QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators
Figure 4 for QAPPA: Quantization-Aware Power, Performance, and Area Modeling of DNN Accelerators
Viaarxiv icon

Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization

Add code
Jul 01, 2019
Figure 1 for Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization
Figure 2 for Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization
Figure 3 for Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization
Figure 4 for Single-Path Mobile AutoML: Efficient ConvNet Design and NAS Hyperparameter Optimization
Viaarxiv icon

ViP: Virtual Pooling for Accelerating CNN-based Image Classification and Object Detection

Add code
Jun 19, 2019
Figure 1 for ViP: Virtual Pooling for Accelerating CNN-based Image Classification and Object Detection
Figure 2 for ViP: Virtual Pooling for Accelerating CNN-based Image Classification and Object Detection
Figure 3 for ViP: Virtual Pooling for Accelerating CNN-based Image Classification and Object Detection
Figure 4 for ViP: Virtual Pooling for Accelerating CNN-based Image Classification and Object Detection
Viaarxiv icon

Single-Path NAS: Device-Aware Efficient ConvNet Design

Add code
May 10, 2019
Figure 1 for Single-Path NAS: Device-Aware Efficient ConvNet Design
Figure 2 for Single-Path NAS: Device-Aware Efficient ConvNet Design
Figure 3 for Single-Path NAS: Device-Aware Efficient ConvNet Design
Viaarxiv icon

LeGR: Filter Pruning via Learned Global Ranking

Add code
Apr 28, 2019
Figure 1 for LeGR: Filter Pruning via Learned Global Ranking
Figure 2 for LeGR: Filter Pruning via Learned Global Ranking
Figure 3 for LeGR: Filter Pruning via Learned Global Ranking
Figure 4 for LeGR: Filter Pruning via Learned Global Ranking
Viaarxiv icon

Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours

Add code
Apr 05, 2019
Figure 1 for Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours
Figure 2 for Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours
Figure 3 for Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours
Figure 4 for Single-Path NAS: Designing Hardware-Efficient ConvNets in less than 4 Hours
Viaarxiv icon

FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference

Add code
Apr 05, 2019
Figure 1 for FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference
Figure 2 for FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference
Figure 3 for FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference
Figure 4 for FLightNNs: Lightweight Quantized Deep Neural Networks for Fast and Accurate Inference
Viaarxiv icon

Regularizing Activation Distribution for Training Binarized Deep Networks

Add code
Apr 04, 2019
Figure 1 for Regularizing Activation Distribution for Training Binarized Deep Networks
Figure 2 for Regularizing Activation Distribution for Training Binarized Deep Networks
Figure 3 for Regularizing Activation Distribution for Training Binarized Deep Networks
Figure 4 for Regularizing Activation Distribution for Training Binarized Deep Networks
Viaarxiv icon