Picture for Chun Jason Xue

Chun Jason Xue

FlexInfer: Breaking Memory Constraint via Flexible and Efficient Offloading for On-Device LLM Inference

Add code
Mar 04, 2025
Viaarxiv icon

CoT-VLM4Tar: Chain-of-Thought Guided Vision-Language Models for Traffic Anomaly Resolution

Add code
Mar 03, 2025
Viaarxiv icon

When Compression Meets Model Compression: Memory-Efficient Double Compression for Large Language Models

Add code
Feb 21, 2025
Viaarxiv icon

EvoP: Robust LLM Inference via Evolutionary Pruning

Add code
Feb 19, 2025
Viaarxiv icon

A$^2$ATS: Retrieval-Based KV Cache Reduction via Windowed Rotary Position Embedding and Query-Aware Vector Quantization

Add code
Feb 18, 2025
Viaarxiv icon

RALAD: Bridging the Real-to-Sim Domain Gap in Autonomous Driving with Retrieval-Augmented Learning

Add code
Jan 21, 2025
Viaarxiv icon

Autoware.Flex: Human-Instructed Dynamically Reconfigurable Autonomous Driving Systems

Add code
Dec 20, 2024
Viaarxiv icon

GeneQuery: A General QA-based Framework for Spatial Gene Expression Predictions from Histology Images

Add code
Nov 27, 2024
Figure 1 for GeneQuery: A General QA-based Framework for Spatial Gene Expression Predictions from Histology Images
Figure 2 for GeneQuery: A General QA-based Framework for Spatial Gene Expression Predictions from Histology Images
Figure 3 for GeneQuery: A General QA-based Framework for Spatial Gene Expression Predictions from Histology Images
Figure 4 for GeneQuery: A General QA-based Framework for Spatial Gene Expression Predictions from Histology Images
Viaarxiv icon

SHAP-CAT: A interpretable multi-modal framework enhancing WSI classification via virtual staining and shapley-value-based multimodal fusion

Add code
Oct 02, 2024
Viaarxiv icon

CHESS: Optimizing LLM Inference via Channel-Wise Thresholding and Selective Sparsification

Add code
Sep 02, 2024
Figure 1 for CHESS: Optimizing LLM Inference via Channel-Wise Thresholding and Selective Sparsification
Figure 2 for CHESS: Optimizing LLM Inference via Channel-Wise Thresholding and Selective Sparsification
Figure 3 for CHESS: Optimizing LLM Inference via Channel-Wise Thresholding and Selective Sparsification
Figure 4 for CHESS: Optimizing LLM Inference via Channel-Wise Thresholding and Selective Sparsification
Viaarxiv icon