Picture for Huihong Shi

Huihong Shi

Celine

TaQ-DiT: Time-aware Quantization for Diffusion Transformers

Add code
Nov 21, 2024
Viaarxiv icon

M$^2$-ViT: Accelerating Hybrid Vision Transformers with Two-Level Mixed Quantization

Add code
Oct 10, 2024
Viaarxiv icon

NASH: Neural Architecture and Accelerator Search for Multiplication-Reduced Hybrid Models

Add code
Sep 07, 2024
Viaarxiv icon

Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration

Add code
Jun 22, 2024
Figure 1 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Figure 2 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Figure 3 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Figure 4 for Unveiling and Harnessing Hidden Attention Sinks: Enhancing Large Language Models without Training through Attention Calibration
Viaarxiv icon

ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization

Add code
Jun 11, 2024
Viaarxiv icon

P$^2$-ViT: Power-of-Two Post-Training Quantization and Acceleration for Fully Quantized Vision Transformer

Add code
May 30, 2024
Viaarxiv icon

Trio-ViT: Post-Training Quantization and Acceleration for Softmax-Free Efficient Vision Transformer

Add code
May 06, 2024
Viaarxiv icon

An FPGA-Based Reconfigurable Accelerator for Convolution-Transformer Hybrid EfficientViT

Add code
Mar 29, 2024
Viaarxiv icon

A Computationally Efficient Neural Video Compression Accelerator Based on a Sparse CNN-Transformer Hybrid Network

Add code
Dec 19, 2023
Viaarxiv icon

S2R: Exploring a Double-Win Transformer-Based Framework for Ideal and Blind Super-Resolution

Add code
Aug 16, 2023
Viaarxiv icon