Picture for Daniel Jannai

Daniel Jannai

Jamba-1.5: Hybrid Transformer-Mamba Models at Scale

Add code
Aug 22, 2024
Figure 1 for Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
Figure 2 for Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
Figure 3 for Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
Figure 4 for Jamba-1.5: Hybrid Transformer-Mamba Models at Scale
Viaarxiv icon

Human or Not? A Gamified Approach to the Turing Test

Add code
May 31, 2023
Viaarxiv icon

Standing on the Shoulders of Giant Frozen Language Models

Add code
Apr 21, 2022
Figure 1 for Standing on the Shoulders of Giant Frozen Language Models
Figure 2 for Standing on the Shoulders of Giant Frozen Language Models
Figure 3 for Standing on the Shoulders of Giant Frozen Language Models
Figure 4 for Standing on the Shoulders of Giant Frozen Language Models
Viaarxiv icon

The Inductive Bias of In-Context Learning: Rethinking Pretraining Example Design

Add code
Oct 25, 2021
Figure 1 for The Inductive Bias of In-Context Learning: Rethinking Pretraining Example Design
Figure 2 for The Inductive Bias of In-Context Learning: Rethinking Pretraining Example Design
Figure 3 for The Inductive Bias of In-Context Learning: Rethinking Pretraining Example Design
Viaarxiv icon

Which transformer architecture fits my data? A vocabulary bottleneck in self-attention

Add code
May 09, 2021
Figure 1 for Which transformer architecture fits my data? A vocabulary bottleneck in self-attention
Figure 2 for Which transformer architecture fits my data? A vocabulary bottleneck in self-attention
Figure 3 for Which transformer architecture fits my data? A vocabulary bottleneck in self-attention
Figure 4 for Which transformer architecture fits my data? A vocabulary bottleneck in self-attention
Viaarxiv icon