Picture for Taiji Suzuki

Taiji Suzuki

On the Comparison between Multi-modal and Single-modal Contrastive Learning

Add code
Nov 05, 2024
Viaarxiv icon

Pretrained transformer efficiently learns low-dimensional target functions in-context

Add code
Nov 04, 2024
Viaarxiv icon

Provably Transformers Harness Multi-Concept Word Semantics for Efficient In-Context Learning

Add code
Nov 04, 2024
Viaarxiv icon

Dimensionality-induced information loss of outliers in deep neural networks

Add code
Oct 29, 2024
Viaarxiv icon

Transformers Provably Solve Parity Efficiently with Chain of Thought

Add code
Oct 11, 2024
Viaarxiv icon

On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent

Add code
Oct 07, 2024
Figure 1 for On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent
Figure 2 for On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent
Figure 3 for On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent
Figure 4 for On the Optimization and Generalization of Two-layer Transformers with Sign Gradient Descent
Viaarxiv icon

Unveil Benign Overfitting for Transformer in Vision: Training Dynamics, Convergence, and Generalization

Add code
Sep 28, 2024
Viaarxiv icon

Transformers are Minimax Optimal Nonparametric In-Context Learners

Add code
Aug 22, 2024
Viaarxiv icon

Learning sum of diverse features: computational hardness and efficient gradient-based training for ridge combinations

Add code
Jun 17, 2024
Viaarxiv icon

Provably Neural Active Learning Succeeds via Prioritizing Perplexing Samples

Add code
Jun 06, 2024
Viaarxiv icon