Picture for Blake Hechtman

Blake Hechtman

eXmY: A Data Type and Technique for Arbitrary Bit Precision Quantization

Add code
May 22, 2024
Figure 1 for eXmY: A Data Type and Technique for Arbitrary Bit Precision Quantization
Figure 2 for eXmY: A Data Type and Technique for Arbitrary Bit Precision Quantization
Figure 3 for eXmY: A Data Type and Technique for Arbitrary Bit Precision Quantization
Figure 4 for eXmY: A Data Type and Technique for Arbitrary Bit Precision Quantization
Viaarxiv icon

Gemini: A Family of Highly Capable Multimodal Models

Add code
Dec 19, 2023
Viaarxiv icon

TPU-KNN: K Nearest Neighbor Search at Peak FLOP/s

Add code
Jun 30, 2022
Figure 1 for TPU-KNN: K Nearest Neighbor Search at Peak FLOP/s
Figure 2 for TPU-KNN: K Nearest Neighbor Search at Peak FLOP/s
Figure 3 for TPU-KNN: K Nearest Neighbor Search at Peak FLOP/s
Figure 4 for TPU-KNN: K Nearest Neighbor Search at Peak FLOP/s
Viaarxiv icon

Unified Scaling Laws for Routed Language Models

Add code
Feb 09, 2022
Figure 1 for Unified Scaling Laws for Routed Language Models
Figure 2 for Unified Scaling Laws for Routed Language Models
Figure 3 for Unified Scaling Laws for Routed Language Models
Figure 4 for Unified Scaling Laws for Routed Language Models
Viaarxiv icon

Scaling Language Models: Methods, Analysis & Insights from Training Gopher

Add code
Dec 08, 2021
Figure 1 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 2 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 3 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 4 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Viaarxiv icon

GSPMD: General and Scalable Parallelization for ML Computation Graphs

Add code
May 10, 2021
Figure 1 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Figure 2 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Figure 3 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Figure 4 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Viaarxiv icon

Scaling Local Self-Attention for Parameter Efficient Visual Backbones

Add code
Mar 30, 2021
Figure 1 for Scaling Local Self-Attention for Parameter Efficient Visual Backbones
Figure 2 for Scaling Local Self-Attention for Parameter Efficient Visual Backbones
Figure 3 for Scaling Local Self-Attention for Parameter Efficient Visual Backbones
Figure 4 for Scaling Local Self-Attention for Parameter Efficient Visual Backbones
Viaarxiv icon

Exploring the limits of Concurrency in ML Training on Google TPUs

Add code
Nov 07, 2020
Figure 1 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 2 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 3 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 4 for Exploring the limits of Concurrency in ML Training on Google TPUs
Viaarxiv icon

Automatic Cross-Replica Sharding of Weight Update in Data-Parallel Training

Add code
Apr 28, 2020
Figure 1 for Automatic Cross-Replica Sharding of Weight Update in Data-Parallel Training
Figure 2 for Automatic Cross-Replica Sharding of Weight Update in Data-Parallel Training
Figure 3 for Automatic Cross-Replica Sharding of Weight Update in Data-Parallel Training
Figure 4 for Automatic Cross-Replica Sharding of Weight Update in Data-Parallel Training
Viaarxiv icon

Scale MLPerf-0.6 models on Google TPU-v3 Pods

Add code
Oct 02, 2019
Figure 1 for Scale MLPerf-0.6 models on Google TPU-v3 Pods
Figure 2 for Scale MLPerf-0.6 models on Google TPU-v3 Pods
Figure 3 for Scale MLPerf-0.6 models on Google TPU-v3 Pods
Figure 4 for Scale MLPerf-0.6 models on Google TPU-v3 Pods
Viaarxiv icon