Picture for Huanrui Yang

Huanrui Yang

SAFER: Sharpness Aware layer-selective Finetuning for Enhanced Robustness in vision transformers

Add code
Jan 02, 2025
Viaarxiv icon

Taming Sensitive Weights : Noise Perturbation Fine-tuning for Robust LLM Quantization

Add code
Dec 08, 2024
Viaarxiv icon

Personalized Multimodal Large Language Models: A Survey

Add code
Dec 03, 2024
Viaarxiv icon

A Survey of Small Language Models

Add code
Oct 25, 2024
Figure 1 for A Survey of Small Language Models
Figure 2 for A Survey of Small Language Models
Figure 3 for A Survey of Small Language Models
Viaarxiv icon

PAT: Pruning-Aware Tuning for Large Language Models

Add code
Aug 27, 2024
Figure 1 for PAT: Pruning-Aware Tuning for Large Language Models
Figure 2 for PAT: Pruning-Aware Tuning for Large Language Models
Figure 3 for PAT: Pruning-Aware Tuning for Large Language Models
Figure 4 for PAT: Pruning-Aware Tuning for Large Language Models
Viaarxiv icon

Criticality Leveraged Adversarial Training (CLAT) for Boosted Performance via Parameter Efficiency

Add code
Aug 19, 2024
Viaarxiv icon

FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models

Add code
Aug 15, 2024
Figure 1 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Figure 2 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Figure 3 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Figure 4 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Viaarxiv icon

Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance

Add code
Jul 17, 2024
Figure 1 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Figure 2 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Figure 3 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Figure 4 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Viaarxiv icon

Fisher-aware Quantization for DETR Detectors with Critical-category Objectives

Add code
Jul 03, 2024
Figure 1 for Fisher-aware Quantization for DETR Detectors with Critical-category Objectives
Figure 2 for Fisher-aware Quantization for DETR Detectors with Critical-category Objectives
Figure 3 for Fisher-aware Quantization for DETR Detectors with Critical-category Objectives
Figure 4 for Fisher-aware Quantization for DETR Detectors with Critical-category Objectives
Viaarxiv icon

Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time Adaptation

Add code
May 26, 2024
Figure 1 for Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time Adaptation
Figure 2 for Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time Adaptation
Figure 3 for Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time Adaptation
Figure 4 for Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time Adaptation
Viaarxiv icon