Picture for Yuan Meng

Yuan Meng

EMS: Adaptive Evict-then-Merge Strategy for Head-wise KV Cache Compression Based on Global-Local Importance

Add code
Dec 11, 2024
Viaarxiv icon

GAQAT: gradient-adaptive quantization-aware training for domain generalization

Add code
Dec 07, 2024
Viaarxiv icon

Mixed Degradation Image Restoration via Local Dynamic Optimization and Conditional Embedding

Add code
Nov 25, 2024
Viaarxiv icon

CATP-LLM: Empowering Large Language Models for Cost-Aware Tool Planning

Add code
Nov 25, 2024
Viaarxiv icon

RTF-Q: Unsupervised domain adaptation based retraining-free quantization network

Add code
Aug 11, 2024
Viaarxiv icon

PRANCE: Joint Token-Optimization and Structural Channel-Pruning for Adaptive ViT Inference

Add code
Jul 06, 2024
Viaarxiv icon

Q-DiT: Accurate Post-Training Quantization for Diffusion Transformers

Add code
Jun 25, 2024
Viaarxiv icon

Towards Lightweight Graph Neural Network Search with Curriculum Graph Sparsification

Add code
Jun 24, 2024
Figure 1 for Towards Lightweight Graph Neural Network Search with Curriculum Graph Sparsification
Figure 2 for Towards Lightweight Graph Neural Network Search with Curriculum Graph Sparsification
Figure 3 for Towards Lightweight Graph Neural Network Search with Curriculum Graph Sparsification
Figure 4 for Towards Lightweight Graph Neural Network Search with Curriculum Graph Sparsification
Viaarxiv icon

One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments

Add code
May 30, 2024
Figure 1 for One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments
Figure 2 for One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments
Figure 3 for One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments
Figure 4 for One QuantLLM for ALL: Fine-tuning Quantized LLMs Once for Efficient Deployments
Viaarxiv icon

LLM-Enhanced Causal Discovery in Temporal Domain from Interventional Data

Add code
Apr 23, 2024
Viaarxiv icon