Picture for Shu Liu

Shu Liu

Fùxì: A Benchmark for Evaluating Language Models on Ancient Chinese Text Understanding and Generation

Add code
Mar 20, 2025
Viaarxiv icon

STEVE: AStep Verification Pipeline for Computer-use Agent Training

Add code
Mar 16, 2025
Viaarxiv icon

The Danger of Overthinking: Examining the Reasoning-Action Dilemma in Agentic Tasks

Add code
Feb 12, 2025
Viaarxiv icon

LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters!

Add code
Feb 11, 2025
Viaarxiv icon

Adaptive Semantic Prompt Caching with VectorQ

Add code
Feb 06, 2025
Viaarxiv icon

Locality-aware Fair Scheduling in LLM Serving

Add code
Jan 24, 2025
Viaarxiv icon

Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition

Add code
Dec 12, 2024
Figure 1 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Figure 2 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Figure 3 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Figure 4 for Lyra: An Efficient and Speech-Centric Framework for Omni-Cognition
Viaarxiv icon

MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs

Add code
Nov 18, 2024
Figure 1 for MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs
Figure 2 for MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs
Figure 3 for MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs
Figure 4 for MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs
Viaarxiv icon

Pie: Pooling CPU Memory for LLM Inference

Add code
Nov 14, 2024
Figure 1 for Pie: Pooling CPU Memory for LLM Inference
Figure 2 for Pie: Pooling CPU Memory for LLM Inference
Figure 3 for Pie: Pooling CPU Memory for LLM Inference
Figure 4 for Pie: Pooling CPU Memory for LLM Inference
Viaarxiv icon

A Natural Primal-Dual Hybrid Gradient Method for Adversarial Neural Network Training on Solving Partial Differential Equations

Add code
Nov 09, 2024
Viaarxiv icon