Picture for Minyi Guo

Minyi Guo

Prism: Mining Task-aware Domains in Non-i.i.d. IMU Data for Flexible User Perception

Add code
Jan 03, 2025
Figure 1 for Prism: Mining Task-aware Domains in Non-i.i.d. IMU Data for Flexible User Perception
Figure 2 for Prism: Mining Task-aware Domains in Non-i.i.d. IMU Data for Flexible User Perception
Figure 3 for Prism: Mining Task-aware Domains in Non-i.i.d. IMU Data for Flexible User Perception
Figure 4 for Prism: Mining Task-aware Domains in Non-i.i.d. IMU Data for Flexible User Perception
Viaarxiv icon

A Survey on Inference Optimization Techniques for Mixture of Experts Models

Add code
Dec 18, 2024
Figure 1 for A Survey on Inference Optimization Techniques for Mixture of Experts Models
Figure 2 for A Survey on Inference Optimization Techniques for Mixture of Experts Models
Figure 3 for A Survey on Inference Optimization Techniques for Mixture of Experts Models
Figure 4 for A Survey on Inference Optimization Techniques for Mixture of Experts Models
Viaarxiv icon

ClusterKV: Manipulating LLM KV Cache in Semantic Space for Recallable Compression

Add code
Dec 04, 2024
Viaarxiv icon

Nimbus: Secure and Efficient Two-Party Inference for Transformers

Add code
Nov 24, 2024
Figure 1 for Nimbus: Secure and Efficient Two-Party Inference for Transformers
Figure 2 for Nimbus: Secure and Efficient Two-Party Inference for Transformers
Figure 3 for Nimbus: Secure and Efficient Two-Party Inference for Transformers
Figure 4 for Nimbus: Secure and Efficient Two-Party Inference for Transformers
Viaarxiv icon

HOBBIT: A Mixed Precision Expert Offloading System for Fast MoE Inference

Add code
Nov 03, 2024
Viaarxiv icon

SparseTem: Boosting the Efficiency of CNN-Based Video Encoders by Exploiting Temporal Continuity

Add code
Oct 28, 2024
Viaarxiv icon

vTensor: Flexible Virtual Tensor Management for Efficient LLM Serving

Add code
Jul 22, 2024
Figure 1 for vTensor: Flexible Virtual Tensor Management for Efficient LLM Serving
Figure 2 for vTensor: Flexible Virtual Tensor Management for Efficient LLM Serving
Figure 3 for vTensor: Flexible Virtual Tensor Management for Efficient LLM Serving
Figure 4 for vTensor: Flexible Virtual Tensor Management for Efficient LLM Serving
Viaarxiv icon

AutoVCoder: A Systematic Framework for Automated Verilog Code Generation using LLMs

Add code
Jul 21, 2024
Figure 1 for AutoVCoder: A Systematic Framework for Automated Verilog Code Generation using LLMs
Figure 2 for AutoVCoder: A Systematic Framework for Automated Verilog Code Generation using LLMs
Figure 3 for AutoVCoder: A Systematic Framework for Automated Verilog Code Generation using LLMs
Figure 4 for AutoVCoder: A Systematic Framework for Automated Verilog Code Generation using LLMs
Viaarxiv icon

SimGen: Simulator-conditioned Driving Scene Generation

Add code
Jun 13, 2024
Figure 1 for SimGen: Simulator-conditioned Driving Scene Generation
Figure 2 for SimGen: Simulator-conditioned Driving Scene Generation
Figure 3 for SimGen: Simulator-conditioned Driving Scene Generation
Figure 4 for SimGen: Simulator-conditioned Driving Scene Generation
Viaarxiv icon

A Codesign of Scheduling and Parallelization for Large Model Training in Heterogeneous Clusters

Add code
Mar 24, 2024
Viaarxiv icon