Picture for Shenggui Li

Shenggui Li

GliDe with a CaPE: A Low-Hassle Method to Accelerate Speculative Decoding

Add code
Feb 03, 2024
Viaarxiv icon

Colossal-Auto: Unified Automation of Parallelization and Activation Checkpoint for Large-scale Models

Add code
Feb 22, 2023
Figure 1 for Colossal-Auto: Unified Automation of Parallelization and Activation Checkpoint for Large-scale Models
Figure 2 for Colossal-Auto: Unified Automation of Parallelization and Activation Checkpoint for Large-scale Models
Figure 3 for Colossal-Auto: Unified Automation of Parallelization and Activation Checkpoint for Large-scale Models
Figure 4 for Colossal-Auto: Unified Automation of Parallelization and Activation Checkpoint for Large-scale Models
Viaarxiv icon

Elixir: Train a Large Language Model on a Small GPU Cluster

Add code
Dec 10, 2022
Viaarxiv icon

EnergonAI: An Inference System for 10-100 Billion Parameter Transformer Models

Add code
Sep 06, 2022
Figure 1 for EnergonAI: An Inference System for 10-100 Billion Parameter Transformer Models
Figure 2 for EnergonAI: An Inference System for 10-100 Billion Parameter Transformer Models
Figure 3 for EnergonAI: An Inference System for 10-100 Billion Parameter Transformer Models
Figure 4 for EnergonAI: An Inference System for 10-100 Billion Parameter Transformer Models
Viaarxiv icon

A Frequency-aware Software Cache for Large Recommendation System Embeddings

Add code
Aug 08, 2022
Figure 1 for A Frequency-aware Software Cache for Large Recommendation System Embeddings
Figure 2 for A Frequency-aware Software Cache for Large Recommendation System Embeddings
Figure 3 for A Frequency-aware Software Cache for Large Recommendation System Embeddings
Figure 4 for A Frequency-aware Software Cache for Large Recommendation System Embeddings
Viaarxiv icon

Sky Computing: Accelerating Geo-distributed Computing in Federated Learning

Add code
Feb 24, 2022
Figure 1 for Sky Computing: Accelerating Geo-distributed Computing in Federated Learning
Figure 2 for Sky Computing: Accelerating Geo-distributed Computing in Federated Learning
Figure 3 for Sky Computing: Accelerating Geo-distributed Computing in Federated Learning
Figure 4 for Sky Computing: Accelerating Geo-distributed Computing in Federated Learning
Viaarxiv icon

PatrickStar: Parallel Training of Pre-trained Models via a Chunk-based Memory Management

Add code
Aug 12, 2021
Figure 1 for PatrickStar: Parallel Training of Pre-trained Models via a Chunk-based Memory Management
Figure 2 for PatrickStar: Parallel Training of Pre-trained Models via a Chunk-based Memory Management
Figure 3 for PatrickStar: Parallel Training of Pre-trained Models via a Chunk-based Memory Management
Figure 4 for PatrickStar: Parallel Training of Pre-trained Models via a Chunk-based Memory Management
Viaarxiv icon

Online Evolutionary Batch Size Orchestration for Scheduling Deep Learning Workloads in GPU Clusters

Add code
Aug 08, 2021
Figure 1 for Online Evolutionary Batch Size Orchestration for Scheduling Deep Learning Workloads in GPU Clusters
Figure 2 for Online Evolutionary Batch Size Orchestration for Scheduling Deep Learning Workloads in GPU Clusters
Figure 3 for Online Evolutionary Batch Size Orchestration for Scheduling Deep Learning Workloads in GPU Clusters
Figure 4 for Online Evolutionary Batch Size Orchestration for Scheduling Deep Learning Workloads in GPU Clusters
Viaarxiv icon

Sequence Parallelism: Long Sequence Training from System Perspective

Add code
May 26, 2021
Figure 1 for Sequence Parallelism: Long Sequence Training from System Perspective
Figure 2 for Sequence Parallelism: Long Sequence Training from System Perspective
Figure 3 for Sequence Parallelism: Long Sequence Training from System Perspective
Figure 4 for Sequence Parallelism: Long Sequence Training from System Perspective
Viaarxiv icon

An Efficient 2D Method for Training Super-Large Deep Learning Models

Add code
Apr 12, 2021
Figure 1 for An Efficient 2D Method for Training Super-Large Deep Learning Models
Figure 2 for An Efficient 2D Method for Training Super-Large Deep Learning Models
Figure 3 for An Efficient 2D Method for Training Super-Large Deep Learning Models
Figure 4 for An Efficient 2D Method for Training Super-Large Deep Learning Models
Viaarxiv icon