Picture for Minlan Yu

Minlan Yu

Harvard University

Minder: Faulty Machine Detection for Large-scale Distributed Model Training

Add code
Nov 04, 2024
Figure 1 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Figure 2 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Figure 3 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Figure 4 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Viaarxiv icon

Federated Learning Clients Clustering with Adaptation to Data Drifts

Add code
Nov 03, 2024
Viaarxiv icon

NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference

Add code
Nov 02, 2024
Figure 1 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Figure 2 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Figure 3 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Figure 4 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Viaarxiv icon

Cora: Accelerating Stateful Network Applications with SmartNICs

Add code
Oct 29, 2024
Figure 1 for Cora: Accelerating Stateful Network Applications with SmartNICs
Figure 2 for Cora: Accelerating Stateful Network Applications with SmartNICs
Figure 3 for Cora: Accelerating Stateful Network Applications with SmartNICs
Figure 4 for Cora: Accelerating Stateful Network Applications with SmartNICs
Viaarxiv icon

Fast Inference for Augmented Large Language Models

Add code
Oct 25, 2024
Figure 1 for Fast Inference for Augmented Large Language Models
Figure 2 for Fast Inference for Augmented Large Language Models
Figure 3 for Fast Inference for Augmented Large Language Models
Figure 4 for Fast Inference for Augmented Large Language Models
Viaarxiv icon

Efficient Inference for Augmented Large Language Models

Add code
Oct 23, 2024
Figure 1 for Efficient Inference for Augmented Large Language Models
Figure 2 for Efficient Inference for Augmented Large Language Models
Figure 3 for Efficient Inference for Augmented Large Language Models
Figure 4 for Efficient Inference for Augmented Large Language Models
Viaarxiv icon

Don't Stop Me Now: Embedding Based Scheduling for LLMs

Add code
Oct 01, 2024
Viaarxiv icon

Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation

Add code
Aug 07, 2024
Viaarxiv icon

HawkVision: Low-Latency Modeless Edge AI Serving

Add code
May 29, 2024
Viaarxiv icon

Carbon Connect: An Ecosystem for Sustainable Computing

Add code
May 22, 2024
Figure 1 for Carbon Connect: An Ecosystem for Sustainable Computing
Figure 2 for Carbon Connect: An Ecosystem for Sustainable Computing
Figure 3 for Carbon Connect: An Ecosystem for Sustainable Computing
Figure 4 for Carbon Connect: An Ecosystem for Sustainable Computing
Viaarxiv icon