Picture for Kai-Chiang Wu

Kai-Chiang Wu

Quamba: A Post-Training Quantization Recipe for Selective State Space Models

Add code
Oct 17, 2024
Viaarxiv icon

ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration

Add code
Sep 15, 2024
Figure 1 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Figure 2 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Figure 3 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Figure 4 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Viaarxiv icon

Palu: Compressing KV-Cache with Low-Rank Projection

Add code
Jul 30, 2024
Viaarxiv icon

FLORA: Fine-grained Low-Rank Architecture Search for Vision Transformer

Add code
Nov 07, 2023
Viaarxiv icon

Q-YOLOP: Quantization-aware You Only Look Once for Panoptic Driving Perception

Add code
Jul 10, 2023
Figure 1 for Q-YOLOP: Quantization-aware You Only Look Once for Panoptic Driving Perception
Figure 2 for Q-YOLOP: Quantization-aware You Only Look Once for Panoptic Driving Perception
Figure 3 for Q-YOLOP: Quantization-aware You Only Look Once for Panoptic Driving Perception
Viaarxiv icon

FOX-NAS: Fast, On-device and Explainable Neural Architecture Search

Add code
Aug 14, 2021
Figure 1 for FOX-NAS: Fast, On-device and Explainable Neural Architecture Search
Figure 2 for FOX-NAS: Fast, On-device and Explainable Neural Architecture Search
Figure 3 for FOX-NAS: Fast, On-device and Explainable Neural Architecture Search
Figure 4 for FOX-NAS: Fast, On-device and Explainable Neural Architecture Search
Viaarxiv icon