Picture for Vikram Saletore

Vikram Saletore

Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case

Add code
Mar 18, 2021
Figure 1 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Figure 2 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Figure 3 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Figure 4 for Reduced Precision Strategies for Deep Learning: A High Energy Physics Generative Adversarial Network Use Case
Viaarxiv icon

Training Multiscale-CNN for Large Microscopy Image Classification in One Hour

Add code
Oct 03, 2019
Figure 1 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Figure 2 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Figure 3 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Figure 4 for Training Multiscale-CNN for Large Microscopy Image Classification in One Hour
Viaarxiv icon

Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model

Add code
Jun 07, 2019
Figure 1 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Figure 2 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Figure 3 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Figure 4 for Efficient 8-Bit Quantization of Transformer Neural Machine Language Translation Model
Viaarxiv icon

Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models

Add code
May 10, 2019
Figure 1 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 2 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 3 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 4 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Viaarxiv icon

Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train

Add code
Nov 15, 2017
Figure 1 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Figure 2 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Figure 3 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Figure 4 for Scale out for large minibatch SGD: Residual network training on ImageNet-1K with improved accuracy and reduced time to train
Viaarxiv icon