Picture for Minlan Yu

Minlan Yu

Harvard University

NetFlowGen: Leveraging Generative Pre-training for Network Traffic Dynamics

Add code
Dec 30, 2024
Viaarxiv icon

TrainMover: Efficient ML Training Live Migration with No Memory Overhead

Add code
Dec 17, 2024
Viaarxiv icon

Minder: Faulty Machine Detection for Large-scale Distributed Model Training

Add code
Nov 04, 2024
Figure 1 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Figure 2 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Figure 3 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Figure 4 for Minder: Faulty Machine Detection for Large-scale Distributed Model Training
Viaarxiv icon

Federated Learning Clients Clustering with Adaptation to Data Drifts

Add code
Nov 03, 2024
Viaarxiv icon

NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference

Add code
Nov 02, 2024
Figure 1 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Figure 2 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Figure 3 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Figure 4 for NEO: Saving GPU Memory Crisis with CPU Offloading for Online LLM Inference
Viaarxiv icon

Cora: Accelerating Stateful Network Applications with SmartNICs

Add code
Oct 29, 2024
Figure 1 for Cora: Accelerating Stateful Network Applications with SmartNICs
Figure 2 for Cora: Accelerating Stateful Network Applications with SmartNICs
Figure 3 for Cora: Accelerating Stateful Network Applications with SmartNICs
Figure 4 for Cora: Accelerating Stateful Network Applications with SmartNICs
Viaarxiv icon

Fast Inference for Augmented Large Language Models

Add code
Oct 25, 2024
Figure 1 for Fast Inference for Augmented Large Language Models
Figure 2 for Fast Inference for Augmented Large Language Models
Figure 3 for Fast Inference for Augmented Large Language Models
Figure 4 for Fast Inference for Augmented Large Language Models
Viaarxiv icon

Efficient Inference for Augmented Large Language Models

Add code
Oct 23, 2024
Figure 1 for Efficient Inference for Augmented Large Language Models
Figure 2 for Efficient Inference for Augmented Large Language Models
Figure 3 for Efficient Inference for Augmented Large Language Models
Figure 4 for Efficient Inference for Augmented Large Language Models
Viaarxiv icon

Don't Stop Me Now: Embedding Based Scheduling for LLMs

Add code
Oct 01, 2024
Figure 1 for Don't Stop Me Now: Embedding Based Scheduling for LLMs
Figure 2 for Don't Stop Me Now: Embedding Based Scheduling for LLMs
Figure 3 for Don't Stop Me Now: Embedding Based Scheduling for LLMs
Figure 4 for Don't Stop Me Now: Embedding Based Scheduling for LLMs
Viaarxiv icon

Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation

Add code
Aug 07, 2024
Figure 1 for Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation
Figure 2 for Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation
Figure 3 for Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation
Figure 4 for Optimus: Accelerating Large-Scale Multi-Modal LLM Training by Bubble Exploitation
Viaarxiv icon