Picture for Srinivas Sridharan

Srinivas Sridharan

LayerDAG: A Layerwise Autoregressive Diffusion Model for Directed Acyclic Graph Generation

Add code
Nov 04, 2024
Viaarxiv icon

Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces

Add code
May 26, 2023
Figure 1 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Figure 2 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Figure 3 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Figure 4 for Chakra: Advancing Performance Benchmarking and Co-design using Standardized Execution Traces
Viaarxiv icon

ASTRA-sim2.0: Modeling Hierarchical Networks and Disaggregated Systems for Large-model Training at Scale

Add code
Mar 24, 2023
Viaarxiv icon

Mystique: Accurate and Scalable Production AI Benchmarks Generation

Add code
Dec 16, 2022
Viaarxiv icon

Impact of RoCE Congestion Control Policies on Distributed Training of DNNs

Add code
Jul 22, 2022
Figure 1 for Impact of RoCE Congestion Control Policies on Distributed Training of DNNs
Figure 2 for Impact of RoCE Congestion Control Policies on Distributed Training of DNNs
Figure 3 for Impact of RoCE Congestion Control Policies on Distributed Training of DNNs
Figure 4 for Impact of RoCE Congestion Control Policies on Distributed Training of DNNs
Viaarxiv icon

Themis: A Network Bandwidth-Aware Collective Scheduling Policy for Distributed Training of DL Models

Add code
Oct 09, 2021
Figure 1 for Themis: A Network Bandwidth-Aware Collective Scheduling Policy for Distributed Training of DL Models
Figure 2 for Themis: A Network Bandwidth-Aware Collective Scheduling Policy for Distributed Training of DL Models
Figure 3 for Themis: A Network Bandwidth-Aware Collective Scheduling Policy for Distributed Training of DL Models
Figure 4 for Themis: A Network Bandwidth-Aware Collective Scheduling Policy for Distributed Training of DL Models
Viaarxiv icon

High-performance, Distributed Training of Large-scale Deep Learning Recommendation Models

Add code
Apr 15, 2021
Figure 1 for High-performance, Distributed Training of Large-scale Deep Learning Recommendation Models
Figure 2 for High-performance, Distributed Training of Large-scale Deep Learning Recommendation Models
Figure 3 for High-performance, Distributed Training of Large-scale Deep Learning Recommendation Models
Figure 4 for High-performance, Distributed Training of Large-scale Deep Learning Recommendation Models
Viaarxiv icon

Automatic Model Parallelism for Deep Neural Networks with Compiler and Hardware Support

Add code
Jun 11, 2019
Figure 1 for Automatic Model Parallelism for Deep Neural Networks with Compiler and Hardware Support
Viaarxiv icon

Mixed Precision Training of Convolutional Neural Networks using Integer Operations

Add code
Feb 23, 2018
Figure 1 for Mixed Precision Training of Convolutional Neural Networks using Integer Operations
Figure 2 for Mixed Precision Training of Convolutional Neural Networks using Integer Operations
Figure 3 for Mixed Precision Training of Convolutional Neural Networks using Integer Operations
Figure 4 for Mixed Precision Training of Convolutional Neural Networks using Integer Operations
Viaarxiv icon

On Scale-out Deep Learning Training for Cloud and HPC

Add code
Jan 24, 2018
Figure 1 for On Scale-out Deep Learning Training for Cloud and HPC
Figure 2 for On Scale-out Deep Learning Training for Cloud and HPC
Viaarxiv icon