Picture for Zhenheng Tang

Zhenheng Tang

One-shot Federated Learning Methods: A Practical Guide

Add code
Feb 13, 2025
Viaarxiv icon

Mediator: Memory-efficient LLM Merging with Less Parameter Conflicts and Uncertainty Based Routing

Add code
Feb 06, 2025
Viaarxiv icon

Can LLMs Maintain Fundamental Abilities under KV Cache Compression?

Add code
Feb 04, 2025
Figure 1 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 2 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 3 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 4 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Viaarxiv icon

FSMoE: A Flexible and Scalable Training System for Sparse Mixture-of-Experts Models

Add code
Jan 18, 2025
Viaarxiv icon

What Limits LLM-based Human Simulation: LLMs or Our Design?

Add code
Jan 15, 2025
Viaarxiv icon

FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion

Add code
Oct 27, 2024
Figure 1 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 2 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 3 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 4 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Viaarxiv icon

Should We Really Edit Language Models? On the Evaluation of Edited Language Models

Add code
Oct 24, 2024
Figure 1 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 2 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 3 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 4 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Viaarxiv icon

ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference

Add code
Oct 23, 2024
Figure 1 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 2 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 3 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 4 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Viaarxiv icon

FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression

Add code
Oct 16, 2024
Figure 1 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 2 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 3 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 4 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Viaarxiv icon

LPZero: Language Model Zero-cost Proxy Search from Zero

Add code
Oct 07, 2024
Figure 1 for LPZero: Language Model Zero-cost Proxy Search from Zero
Figure 2 for LPZero: Language Model Zero-cost Proxy Search from Zero
Figure 3 for LPZero: Language Model Zero-cost Proxy Search from Zero
Figure 4 for LPZero: Language Model Zero-cost Proxy Search from Zero
Viaarxiv icon