Picture for Caglar Gulcehre

Caglar Gulcehre

The Effect of Scheduling and Preemption on the Efficiency of LLM Inference Serving

Add code
Nov 12, 2024
Viaarxiv icon

Unpacking SDXL Turbo: Interpreting Text-to-Image Models with Sparse Autoencoders

Add code
Oct 28, 2024
Viaarxiv icon

Beyond Autoregression: Fast LLMs via Self-Distillation Through Time

Add code
Oct 28, 2024
Viaarxiv icon

SIKeD: Self-guided Iterative Knowledge Distillation for mathematical reasoning

Add code
Oct 24, 2024
Viaarxiv icon

The Role of Deep Learning Regularizations on Actors in Offline RL

Add code
Sep 11, 2024
Viaarxiv icon

In Search for Architectures and Loss Functions in Multi-Objective Reinforcement Learning

Add code
Jul 23, 2024
Viaarxiv icon

Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis

Add code
Jul 13, 2024
Figure 1 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 2 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 3 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Figure 4 for Investigating Low-Rank Training in Transformer Language Models: Efficiency and Scaling Analysis
Viaarxiv icon

HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context

Add code
Jul 12, 2024
Figure 1 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Figure 2 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Figure 3 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Figure 4 for HiPPO-Prophecy: State-Space Models can Provably Learn Dynamical Systems in Context
Viaarxiv icon

Self-Recognition in Language Models

Add code
Jul 09, 2024
Viaarxiv icon

Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers

Add code
Jun 24, 2024
Viaarxiv icon