Picture for Chi-Chih Chang

Chi-Chih Chang

Quamba: A Post-Training Quantization Recipe for Selective State Space Models

Add code
Oct 17, 2024
Viaarxiv icon

ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration

Add code
Sep 15, 2024
Figure 1 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Figure 2 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Figure 3 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Figure 4 for ELSA: Exploiting Layer-wise N:M Sparsity for Vision Transformer Acceleration
Viaarxiv icon

Palu: Compressing KV-Cache with Low-Rank Projection

Add code
Jul 30, 2024
Viaarxiv icon

FLORA: Fine-grained Low-Rank Architecture Search for Vision Transformer

Add code
Nov 07, 2023
Viaarxiv icon

Q-YOLOP: Quantization-aware You Only Look Once for Panoptic Driving Perception

Add code
Jul 10, 2023
Viaarxiv icon