Picture for Ali Jadbabaie

Ali Jadbabaie

Fast Tensor Completion via Approximate Richardson Iteration

Add code
Feb 13, 2025
Viaarxiv icon

On the Emergence of Position Bias in Transformers

Add code
Feb 04, 2025
Viaarxiv icon

In-Context Learning of Polynomial Kernel Regression in Transformers with GLU Layers

Add code
Jan 30, 2025
Viaarxiv icon

GraphHash: Graph Clustering Enables Parameter Efficiency in Recommender Systems

Add code
Dec 23, 2024
Viaarxiv icon

Improved Sample Complexity of Imitation Learning for Barrier Model Predictive Control

Add code
Oct 01, 2024
Viaarxiv icon

Residual Connections and Normalization Can Provably Prevent Oversmoothing in GNNs

Add code
Jun 05, 2024
Viaarxiv icon

On the Role of Attention Masks and LayerNorm in Transformers

Add code
May 29, 2024
Figure 1 for On the Role of Attention Masks and LayerNorm in Transformers
Figure 2 for On the Role of Attention Masks and LayerNorm in Transformers
Figure 3 for On the Role of Attention Masks and LayerNorm in Transformers
Figure 4 for On the Role of Attention Masks and LayerNorm in Transformers
Viaarxiv icon

A least-square method for non-asymptotic identification in linear switching control

Add code
Apr 11, 2024
Viaarxiv icon

Belief Samples Are All You Need For Social Learning

Add code
Mar 25, 2024
Viaarxiv icon

Linear attention is (maybe) all you need (to understand transformer optimization)

Add code
Oct 02, 2023
Figure 1 for Linear attention is (maybe) all you need (to understand transformer optimization)
Figure 2 for Linear attention is (maybe) all you need (to understand transformer optimization)
Figure 3 for Linear attention is (maybe) all you need (to understand transformer optimization)
Figure 4 for Linear attention is (maybe) all you need (to understand transformer optimization)
Viaarxiv icon