Picture for Xiaowen Chu

Xiaowen Chu

FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion

Add code
Oct 27, 2024
Figure 1 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 2 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 3 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 4 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Viaarxiv icon

Should We Really Edit Language Models? On the Evaluation of Edited Language Models

Add code
Oct 24, 2024
Figure 1 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 2 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 3 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 4 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Viaarxiv icon

ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference

Add code
Oct 23, 2024
Figure 1 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 2 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 3 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 4 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Viaarxiv icon

FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression

Add code
Oct 16, 2024
Figure 1 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 2 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 3 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 4 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Viaarxiv icon

LPZero: Language Model Zero-cost Proxy Search from Zero

Add code
Oct 07, 2024
Figure 1 for LPZero: Language Model Zero-cost Proxy Search from Zero
Figure 2 for LPZero: Language Model Zero-cost Proxy Search from Zero
Figure 3 for LPZero: Language Model Zero-cost Proxy Search from Zero
Figure 4 for LPZero: Language Model Zero-cost Proxy Search from Zero
Viaarxiv icon

LongGenBench: Long-context Generation Benchmark

Add code
Oct 05, 2024
Viaarxiv icon

From Words to Wheels: Automated Style-Customized Policy Generation for Autonomous Driving

Add code
Sep 18, 2024
Figure 1 for From Words to Wheels: Automated Style-Customized Policy Generation for Autonomous Driving
Figure 2 for From Words to Wheels: Automated Style-Customized Policy Generation for Autonomous Driving
Figure 3 for From Words to Wheels: Automated Style-Customized Policy Generation for Autonomous Driving
Figure 4 for From Words to Wheels: Automated Style-Customized Policy Generation for Autonomous Driving
Viaarxiv icon

Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning

Add code
Aug 27, 2024
Figure 1 for Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning
Figure 2 for Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning
Figure 3 for Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning
Figure 4 for Bandwidth-Aware and Overlap-Weighted Compression for Communication-Efficient Federated Learning
Viaarxiv icon

Asteroid: Resource-Efficient Hybrid Pipeline Parallelism for Collaborative DNN Training on Heterogeneous Edge Devices

Add code
Aug 15, 2024
Figure 1 for Asteroid: Resource-Efficient Hybrid Pipeline Parallelism for Collaborative DNN Training on Heterogeneous Edge Devices
Figure 2 for Asteroid: Resource-Efficient Hybrid Pipeline Parallelism for Collaborative DNN Training on Heterogeneous Edge Devices
Figure 3 for Asteroid: Resource-Efficient Hybrid Pipeline Parallelism for Collaborative DNN Training on Heterogeneous Edge Devices
Figure 4 for Asteroid: Resource-Efficient Hybrid Pipeline Parallelism for Collaborative DNN Training on Heterogeneous Edge Devices
Viaarxiv icon

STBLLM: Breaking the 1-Bit Barrier with Structured Binary LLMs

Add code
Aug 03, 2024
Viaarxiv icon