Picture for Yuanchun Li

Yuanchun Li

Institute for AI Industry Research, Shanghai AI Laboratory, Shanghai, China

A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage

Add code
Sep 06, 2024
Figure 1 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Figure 2 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Figure 3 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Figure 4 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Viaarxiv icon

LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design

Add code
May 28, 2024
Viaarxiv icon

LlamaTouch: A Faithful and Scalable Testbed for Mobile UI Automation Task Evaluation

Add code
Apr 12, 2024
Viaarxiv icon

Exploring the Impact of In-Browser Deep Learning Inference on Quality of User Experience and Performance

Add code
Feb 08, 2024
Viaarxiv icon

A Survey of Resource-efficient LLM and Multimodal Foundation Models

Add code
Jan 16, 2024
Figure 1 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Figure 2 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Figure 3 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Figure 4 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Viaarxiv icon

Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security

Add code
Jan 10, 2024
Figure 1 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Figure 2 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Figure 3 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Figure 4 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Viaarxiv icon

Accelerating In-Browser Deep Learning Inference on Diverse Edge Clients through Just-in-Time Kernel Optimizations

Add code
Sep 16, 2023
Viaarxiv icon

Empowering LLM to use Smartphone for Intelligent Task Automation

Add code
Sep 09, 2023
Viaarxiv icon

Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping

Add code
Aug 29, 2023
Figure 1 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Figure 2 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Figure 3 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Figure 4 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Viaarxiv icon

Generative Model for Models: Rapid DNN Customization for Diverse Tasks and Resource Constraints

Add code
Aug 29, 2023
Viaarxiv icon