Picture for Shiyun Wei

Shiyun Wei

LLMCad: Fast and Scalable On-device Large Language Model Inference

Add code
Sep 08, 2023
Viaarxiv icon

EdgeMoE: Fast On-Device Inference of MoE-based Large Language Models

Add code
Aug 28, 2023
Figure 1 for EdgeMoE: Fast On-Device Inference of MoE-based Large Language Models
Figure 2 for EdgeMoE: Fast On-Device Inference of MoE-based Large Language Models
Figure 3 for EdgeMoE: Fast On-Device Inference of MoE-based Large Language Models
Figure 4 for EdgeMoE: Fast On-Device Inference of MoE-based Large Language Models
Viaarxiv icon