Picture for Alexander Kozachinskiy

Alexander Kozachinskiy

NoPE: The Counting Power of Transformers with No Positional Encodings

Add code
May 16, 2025
Viaarxiv icon

Continuity and Isolation Lead to Doubts or Dilemmas in Large Language Models

Add code
May 15, 2025
Viaarxiv icon

Strassen Attention: Unlocking Compositional Abilities in Transformers Based on a New Lower Bound Method

Add code
Jan 31, 2025
Figure 1 for Strassen Attention: Unlocking Compositional Abilities in Transformers Based on a New Lower Bound Method
Figure 2 for Strassen Attention: Unlocking Compositional Abilities in Transformers Based on a New Lower Bound Method
Figure 3 for Strassen Attention: Unlocking Compositional Abilities in Transformers Based on a New Lower Bound Method
Figure 4 for Strassen Attention: Unlocking Compositional Abilities in Transformers Based on a New Lower Bound Method
Viaarxiv icon

Explaining k-Nearest Neighbors: Abductive and Counterfactual Explanations

Add code
Jan 10, 2025
Viaarxiv icon

A completely uniform transformer for parity

Add code
Jan 05, 2025
Viaarxiv icon

Lower bounds on transformers with infinite precision

Add code
Dec 28, 2024
Viaarxiv icon

Effective Littlestone Dimension

Add code
Nov 22, 2024
Viaarxiv icon

On dimensionality of feature vectors in MPNNs

Add code
Feb 14, 2024
Viaarxiv icon

Logical Languages Accepted by Transformer Encoders with Hard Attention

Add code
Oct 05, 2023
Viaarxiv icon

Simple online learning with consistency oracle

Add code
Aug 15, 2023
Viaarxiv icon