Picture for Anselm Levskaya

Anselm Levskaya

Gemini 1.5: Unlocking multimodal understanding across millions of tokens of context

Add code
Mar 08, 2024
Viaarxiv icon

Gemini: A Family of Highly Capable Multimodal Models

Add code
Dec 19, 2023
Viaarxiv icon

Efficiently Scaling Transformer Inference

Add code
Nov 09, 2022
Viaarxiv icon

PaLM: Scaling Language Modeling with Pathways

Add code
Apr 19, 2022
Figure 1 for PaLM: Scaling Language Modeling with Pathways
Figure 2 for PaLM: Scaling Language Modeling with Pathways
Figure 3 for PaLM: Scaling Language Modeling with Pathways
Figure 4 for PaLM: Scaling Language Modeling with Pathways
Viaarxiv icon

Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$

Add code
Mar 31, 2022
Figure 1 for Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$
Figure 2 for Scaling Up Models and Data with $\texttt{t5x}$ and $\texttt{seqio}$
Viaarxiv icon

Exploring the limits of Concurrency in ML Training on Google TPUs

Add code
Nov 07, 2020
Figure 1 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 2 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 3 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 4 for Exploring the limits of Concurrency in ML Training on Google TPUs
Viaarxiv icon

Reformer: The Efficient Transformer

Add code
Feb 18, 2020
Figure 1 for Reformer: The Efficient Transformer
Figure 2 for Reformer: The Efficient Transformer
Figure 3 for Reformer: The Efficient Transformer
Figure 4 for Reformer: The Efficient Transformer
Viaarxiv icon

Stand-Alone Self-Attention in Vision Models

Add code
Jun 13, 2019
Figure 1 for Stand-Alone Self-Attention in Vision Models
Figure 2 for Stand-Alone Self-Attention in Vision Models
Figure 3 for Stand-Alone Self-Attention in Vision Models
Figure 4 for Stand-Alone Self-Attention in Vision Models
Viaarxiv icon