Picture for Hugo Touvron

Hugo Touvron

Jack

The Llama 3 Herd of Models

Add code
Jul 31, 2024
Viaarxiv icon

Automatic Data Curation for Self-Supervised Learning: A Clustering-Based Approach

Add code
May 24, 2024
Viaarxiv icon

Code Llama: Open Foundation Models for Code

Add code
Aug 25, 2023
Figure 1 for Code Llama: Open Foundation Models for Code
Figure 2 for Code Llama: Open Foundation Models for Code
Figure 3 for Code Llama: Open Foundation Models for Code
Figure 4 for Code Llama: Open Foundation Models for Code
Viaarxiv icon

Llama 2: Open Foundation and Fine-Tuned Chat Models

Add code
Jul 19, 2023
Figure 1 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 2 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 3 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Figure 4 for Llama 2: Open Foundation and Fine-Tuned Chat Models
Viaarxiv icon

LLaMA: Open and Efficient Foundation Language Models

Add code
Feb 27, 2023
Viaarxiv icon

Co-training $2^L$ Submodels for Visual Recognition

Add code
Dec 09, 2022
Viaarxiv icon

DeiT III: Revenge of the ViT

Add code
Apr 14, 2022
Figure 1 for DeiT III: Revenge of the ViT
Figure 2 for DeiT III: Revenge of the ViT
Figure 3 for DeiT III: Revenge of the ViT
Figure 4 for DeiT III: Revenge of the ViT
Viaarxiv icon

Three things everyone should know about Vision Transformers

Add code
Mar 18, 2022
Figure 1 for Three things everyone should know about Vision Transformers
Figure 2 for Three things everyone should know about Vision Transformers
Figure 3 for Three things everyone should know about Vision Transformers
Figure 4 for Three things everyone should know about Vision Transformers
Viaarxiv icon

Augmenting Convolutional networks with attention-based aggregation

Add code
Dec 27, 2021
Figure 1 for Augmenting Convolutional networks with attention-based aggregation
Figure 2 for Augmenting Convolutional networks with attention-based aggregation
Figure 3 for Augmenting Convolutional networks with attention-based aggregation
Figure 4 for Augmenting Convolutional networks with attention-based aggregation
Viaarxiv icon

Are Large-scale Datasets Necessary for Self-Supervised Pre-training?

Add code
Dec 20, 2021
Figure 1 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Figure 2 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Figure 3 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Figure 4 for Are Large-scale Datasets Necessary for Self-Supervised Pre-training?
Viaarxiv icon