Picture for Samet Oymak

Samet Oymak

Selective Attention: Enhancing Transformer through Principled Context Control

Add code
Nov 19, 2024
Viaarxiv icon

Retrieval Augmented Time Series Forecasting

Add code
Nov 12, 2024
Figure 1 for Retrieval Augmented Time Series Forecasting
Figure 2 for Retrieval Augmented Time Series Forecasting
Figure 3 for Retrieval Augmented Time Series Forecasting
Figure 4 for Retrieval Augmented Time Series Forecasting
Viaarxiv icon

High-dimensional Analysis of Knowledge Distillation: Weak-to-Strong Generalization and Scaling Laws

Add code
Oct 24, 2024
Viaarxiv icon

Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition

Add code
Oct 08, 2024
Figure 1 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 2 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 3 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Figure 4 for Everything Everywhere All at Once: LLMs can In-Context Learn Multiple Tasks in Superposition
Viaarxiv icon

Fine-grained Analysis of In-context Linear Estimation: Data, Architecture, and Beyond

Add code
Jul 13, 2024
Viaarxiv icon

On the Power of Convolution Augmented Transformer

Add code
Jul 08, 2024
Viaarxiv icon

TREACLE: Thrifty Reasoning via Context-Aware LLM and Prompt Selection

Add code
Apr 17, 2024
Viaarxiv icon

Mechanics of Next Token Prediction with Self-Attention

Add code
Mar 12, 2024
Figure 1 for Mechanics of Next Token Prediction with Self-Attention
Figure 2 for Mechanics of Next Token Prediction with Self-Attention
Figure 3 for Mechanics of Next Token Prediction with Self-Attention
Figure 4 for Mechanics of Next Token Prediction with Self-Attention
Viaarxiv icon

From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers

Add code
Feb 21, 2024
Figure 1 for From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers
Figure 2 for From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers
Figure 3 for From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers
Figure 4 for From Self-Attention to Markov Models: Unveiling the Dynamics of Generative Transformers
Viaarxiv icon

FLASH: Federated Learning Across Simultaneous Heterogeneities

Add code
Feb 13, 2024
Viaarxiv icon