Picture for Xiaowen Chu

Xiaowen Chu

Perovskite-LLM: Knowledge-Enhanced Large Language Models for Perovskite Solar Cell Research

Add code
Feb 18, 2025
Viaarxiv icon

SphereFusion: Efficient Panorama Depth Estimation via Gated Fusion

Add code
Feb 09, 2025
Viaarxiv icon

Mediator: Memory-efficient LLM Merging with Less Parameter Conflicts and Uncertainty Based Routing

Add code
Feb 06, 2025
Viaarxiv icon

Can LLMs Maintain Fundamental Abilities under KV Cache Compression?

Add code
Feb 04, 2025
Figure 1 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 2 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 3 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 4 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Viaarxiv icon

FSMoE: A Flexible and Scalable Training System for Sparse Mixture-of-Experts Models

Add code
Jan 18, 2025
Viaarxiv icon

Accelerating AIGC Services with Latent Action Diffusion Scheduling in Edge Networks

Add code
Dec 24, 2024
Figure 1 for Accelerating AIGC Services with Latent Action Diffusion Scheduling in Edge Networks
Figure 2 for Accelerating AIGC Services with Latent Action Diffusion Scheduling in Edge Networks
Figure 3 for Accelerating AIGC Services with Latent Action Diffusion Scheduling in Edge Networks
Figure 4 for Accelerating AIGC Services with Latent Action Diffusion Scheduling in Edge Networks
Viaarxiv icon

FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion

Add code
Oct 27, 2024
Figure 1 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 2 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 3 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Figure 4 for FuseFL: One-Shot Federated Learning through the Lens of Causality with Progressive Model Fusion
Viaarxiv icon

Should We Really Edit Language Models? On the Evaluation of Edited Language Models

Add code
Oct 24, 2024
Figure 1 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 2 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 3 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Figure 4 for Should We Really Edit Language Models? On the Evaluation of Edited Language Models
Viaarxiv icon

ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference

Add code
Oct 23, 2024
Figure 1 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 2 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 3 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Figure 4 for ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference
Viaarxiv icon

FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression

Add code
Oct 16, 2024
Figure 1 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 2 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 3 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Figure 4 for FusionLLM: A Decentralized LLM Training System on Geo-distributed GPUs with Adaptive Compression
Viaarxiv icon