Picture for Zhenmei Shi

Zhenmei Shi

On the Computational Capability of Graph Neural Networks: A Circuit Complexity Bound Perspective

Add code
Jan 11, 2025
Viaarxiv icon

On Computational Limits and Provably Efficient Criteria of Visual Autoregressive Models: A Fine-Grained Complexity Analysis

Add code
Jan 08, 2025
Viaarxiv icon

Circuit Complexity Bounds for Visual Autoregressive Model

Add code
Jan 08, 2025
Viaarxiv icon

Theoretical Constraints on the Expressive Power of $\mathsf{RoPE}$-based Tensor Attention Transformers

Add code
Dec 23, 2024
Viaarxiv icon

Fast Gradient Computation for RoPE Attention in Almost Linear Time

Add code
Dec 23, 2024
Viaarxiv icon

The Computational Limits of State-Space Models and Mamba via the Lens of Circuit Complexity

Add code
Dec 09, 2024
Viaarxiv icon

Curse of Attention: A Kernel-Based Perspective for Why Transformers Fail to Generalize on Time Series Forecasting and Beyond

Add code
Dec 08, 2024
Viaarxiv icon

On the Expressive Power of Modern Hopfield Networks

Add code
Dec 07, 2024
Viaarxiv icon

Circuit Complexity Bounds for RoPE-based Transformer Architecture

Add code
Nov 12, 2024
Viaarxiv icon

Bypassing the Exponential Dependency: Looped Transformers Efficiently Learn In-context by Multi-step Gradient Descent

Add code
Oct 15, 2024
Figure 1 for Bypassing the Exponential Dependency: Looped Transformers Efficiently Learn In-context by Multi-step Gradient Descent
Viaarxiv icon