Picture for Ruisi Cai

Ruisi Cai

Read-ME: Refactorizing LLMs as Router-Decoupled Mixture of Experts with System Co-Design

Add code
Oct 24, 2024
Figure 1 for Read-ME: Refactorizing LLMs as Router-Decoupled Mixture of Experts with System Co-Design
Figure 2 for Read-ME: Refactorizing LLMs as Router-Decoupled Mixture of Experts with System Co-Design
Figure 3 for Read-ME: Refactorizing LLMs as Router-Decoupled Mixture of Experts with System Co-Design
Figure 4 for Read-ME: Refactorizing LLMs as Router-Decoupled Mixture of Experts with System Co-Design
Viaarxiv icon

Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild

Add code
Oct 07, 2024
Figure 1 for Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild
Figure 2 for Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild
Figure 3 for Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild
Figure 4 for Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the Wild
Viaarxiv icon

Flextron: Many-in-One Flexible Large Language Model

Add code
Jun 11, 2024
Figure 1 for Flextron: Many-in-One Flexible Large Language Model
Figure 2 for Flextron: Many-in-One Flexible Large Language Model
Figure 3 for Flextron: Many-in-One Flexible Large Language Model
Figure 4 for Flextron: Many-in-One Flexible Large Language Model
Viaarxiv icon

LoCoCo: Dropping In Convolutions for Long Context Compression

Add code
Jun 08, 2024
Viaarxiv icon

Robust Mixture-of-Expert Training for Convolutional Neural Networks

Add code
Aug 19, 2023
Viaarxiv icon

H$_2$O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models

Add code
Jul 19, 2023
Viaarxiv icon

Robust Weight Signatures: Gaining Robustness as Easy as Patching Weights?

Add code
Feb 24, 2023
Viaarxiv icon