Picture for Can Karakus

Can Karakus

MADA: Meta-Adaptive Optimizers through hyper-gradient Descent

Add code
Jan 17, 2024
Viaarxiv icon

Amazon SageMaker Model Parallelism: A General and Flexible Framework for Large Model Training

Add code
Nov 10, 2021
Figure 1 for Amazon SageMaker Model Parallelism: A General and Flexible Framework for Large Model Training
Figure 2 for Amazon SageMaker Model Parallelism: A General and Flexible Framework for Large Model Training
Figure 3 for Amazon SageMaker Model Parallelism: A General and Flexible Framework for Large Model Training
Figure 4 for Amazon SageMaker Model Parallelism: A General and Flexible Framework for Large Model Training
Viaarxiv icon

Qsparse-local-SGD: Distributed SGD with Quantization, Sparsification, and Local Computations

Add code
Jun 06, 2019
Figure 1 for Qsparse-local-SGD: Distributed SGD with Quantization, Sparsification, and Local Computations
Figure 2 for Qsparse-local-SGD: Distributed SGD with Quantization, Sparsification, and Local Computations
Figure 3 for Qsparse-local-SGD: Distributed SGD with Quantization, Sparsification, and Local Computations
Viaarxiv icon

Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models

Add code
May 10, 2019
Figure 1 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 2 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 3 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Figure 4 for Densifying Assumed-sparse Tensors: Improving Memory Efficiency and MPI Collective Performance during Tensor Accumulation for Parallelized Training of Neural Machine Translation Models
Viaarxiv icon

Differentially Private Consensus-Based Distributed Optimization

Add code
Mar 19, 2019
Figure 1 for Differentially Private Consensus-Based Distributed Optimization
Figure 2 for Differentially Private Consensus-Based Distributed Optimization
Figure 3 for Differentially Private Consensus-Based Distributed Optimization
Figure 4 for Differentially Private Consensus-Based Distributed Optimization
Viaarxiv icon

Privacy-Utility Trade-off of Linear Regression under Random Projections and Additive Noise

Add code
Feb 13, 2019
Figure 1 for Privacy-Utility Trade-off of Linear Regression under Random Projections and Additive Noise
Figure 2 for Privacy-Utility Trade-off of Linear Regression under Random Projections and Additive Noise
Figure 3 for Privacy-Utility Trade-off of Linear Regression under Random Projections and Additive Noise
Viaarxiv icon

Redundancy Techniques for Straggler Mitigation in Distributed Optimization and Learning

Add code
Mar 14, 2018
Figure 1 for Redundancy Techniques for Straggler Mitigation in Distributed Optimization and Learning
Figure 2 for Redundancy Techniques for Straggler Mitigation in Distributed Optimization and Learning
Figure 3 for Redundancy Techniques for Straggler Mitigation in Distributed Optimization and Learning
Figure 4 for Redundancy Techniques for Straggler Mitigation in Distributed Optimization and Learning
Viaarxiv icon

Straggler Mitigation in Distributed Optimization Through Data Encoding

Add code
Jan 22, 2018
Figure 1 for Straggler Mitigation in Distributed Optimization Through Data Encoding
Figure 2 for Straggler Mitigation in Distributed Optimization Through Data Encoding
Figure 3 for Straggler Mitigation in Distributed Optimization Through Data Encoding
Figure 4 for Straggler Mitigation in Distributed Optimization Through Data Encoding
Viaarxiv icon