Picture for Yuanchun Li

Yuanchun Li

Institute for AI Industry Research, Shanghai AI Laboratory, Shanghai, China

A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage

Add code
Sep 06, 2024
Figure 1 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Figure 2 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Figure 3 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Figure 4 for A First Look At Efficient And Secure On-Device LLM Inference Against KV Leakage
Viaarxiv icon

LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design

Add code
May 28, 2024
Viaarxiv icon

LlamaTouch: A Faithful and Scalable Testbed for Mobile UI Automation Task Evaluation

Add code
Apr 12, 2024
Viaarxiv icon

Exploring the Impact of In-Browser Deep Learning Inference on Quality of User Experience and Performance

Add code
Feb 08, 2024
Figure 1 for Exploring the Impact of In-Browser Deep Learning Inference on Quality of User Experience and Performance
Figure 2 for Exploring the Impact of In-Browser Deep Learning Inference on Quality of User Experience and Performance
Figure 3 for Exploring the Impact of In-Browser Deep Learning Inference on Quality of User Experience and Performance
Figure 4 for Exploring the Impact of In-Browser Deep Learning Inference on Quality of User Experience and Performance
Viaarxiv icon

A Survey of Resource-efficient LLM and Multimodal Foundation Models

Add code
Jan 16, 2024
Figure 1 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Figure 2 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Figure 3 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Figure 4 for A Survey of Resource-efficient LLM and Multimodal Foundation Models
Viaarxiv icon

Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security

Add code
Jan 10, 2024
Figure 1 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Figure 2 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Figure 3 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Figure 4 for Personal LLM Agents: Insights and Survey about the Capability, Efficiency and Security
Viaarxiv icon

Accelerating In-Browser Deep Learning Inference on Diverse Edge Clients through Just-in-Time Kernel Optimizations

Add code
Sep 16, 2023
Figure 1 for Accelerating In-Browser Deep Learning Inference on Diverse Edge Clients through Just-in-Time Kernel Optimizations
Figure 2 for Accelerating In-Browser Deep Learning Inference on Diverse Edge Clients through Just-in-Time Kernel Optimizations
Figure 3 for Accelerating In-Browser Deep Learning Inference on Diverse Edge Clients through Just-in-Time Kernel Optimizations
Figure 4 for Accelerating In-Browser Deep Learning Inference on Diverse Edge Clients through Just-in-Time Kernel Optimizations
Viaarxiv icon

Empowering LLM to use Smartphone for Intelligent Task Automation

Add code
Sep 09, 2023
Viaarxiv icon

Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping

Add code
Aug 29, 2023
Figure 1 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Figure 2 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Figure 3 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Figure 4 for Serving MoE Models on Resource-constrained Edge Devices via Dynamic Expert Swapping
Viaarxiv icon

Generative Model for Models: Rapid DNN Customization for Diverse Tasks and Resource Constraints

Add code
Aug 29, 2023
Viaarxiv icon