Picture for Yufa Zhou

Yufa Zhou

Numerical Pruning for Efficient Autoregressive Models

Add code
Dec 17, 2024
Viaarxiv icon

LazyDiT: Lazy Learning for the Acceleration of Diffusion Transformers

Add code
Dec 17, 2024
Viaarxiv icon

Beyond Linear Approximations: A Novel Pruning Approach for Attention Matrix

Add code
Oct 15, 2024
Viaarxiv icon

Fine-grained Attention I/O Complexity: Comprehensive Analysis for Backward Passes

Add code
Oct 12, 2024
Viaarxiv icon

Looped ReLU MLPs May Be All You Need as Practical Programmable Computers

Add code
Oct 12, 2024
Viaarxiv icon

Multi-Layer Transformers Gradient Can be Approximated in Almost Linear Time

Add code
Aug 23, 2024
Viaarxiv icon

Differential Privacy of Cross-Attention with Provable Guarantee

Add code
Jul 20, 2024
Figure 1 for Differential Privacy of Cross-Attention with Provable Guarantee
Viaarxiv icon

Tensor Attention Training: Provably Efficient Learning of Higher-order Transformers

Add code
May 26, 2024
Viaarxiv icon

Unraveling the Smoothness Properties of Diffusion Models: A Gaussian Mixture Perspective

Add code
May 26, 2024
Viaarxiv icon