Picture for Zhenyu Bai

Zhenyu Bai

Matterhorn: Efficient Analog Sparse Spiking Transformer Architecture with Masked Time-To-First-Spike Encoding

Add code
Jan 30, 2026
Viaarxiv icon

SWAT: Scalable and Efficient Window Attention-based Transformers Acceleration on FPGAs

Add code
May 27, 2024
Figure 1 for SWAT: Scalable and Efficient Window Attention-based Transformers Acceleration on FPGAs
Figure 2 for SWAT: Scalable and Efficient Window Attention-based Transformers Acceleration on FPGAs
Figure 3 for SWAT: Scalable and Efficient Window Attention-based Transformers Acceleration on FPGAs
Figure 4 for SWAT: Scalable and Efficient Window Attention-based Transformers Acceleration on FPGAs
Viaarxiv icon