Picture for Fartash Faghri

Fartash Faghri

FocalLens: Instruction Tuning Enables Zero-Shot Conditional Image Representations

Add code
Apr 11, 2025
Viaarxiv icon

TiC-LM: A Web-Scale Benchmark for Time-Continual LLM Pretraining

Add code
Apr 02, 2025
Viaarxiv icon

FastVLM: Efficient Vision Encoding for Vision Language Models

Add code
Dec 17, 2024
Figure 1 for FastVLM: Efficient Vision Encoding for Vision Language Models
Figure 2 for FastVLM: Efficient Vision Encoding for Vision Language Models
Figure 3 for FastVLM: Efficient Vision Encoding for Vision Language Models
Figure 4 for FastVLM: Efficient Vision Encoding for Vision Language Models
Viaarxiv icon

Computational Bottlenecks of Training Small-scale Large Language Models

Add code
Oct 25, 2024
Figure 1 for Computational Bottlenecks of Training Small-scale Large Language Models
Figure 2 for Computational Bottlenecks of Training Small-scale Large Language Models
Figure 3 for Computational Bottlenecks of Training Small-scale Large Language Models
Figure 4 for Computational Bottlenecks of Training Small-scale Large Language Models
Viaarxiv icon

MUSCLE: A Model Update Strategy for Compatible LLM Evolution

Add code
Jul 12, 2024
Figure 1 for MUSCLE: A Model Update Strategy for Compatible LLM Evolution
Figure 2 for MUSCLE: A Model Update Strategy for Compatible LLM Evolution
Figure 3 for MUSCLE: A Model Update Strategy for Compatible LLM Evolution
Figure 4 for MUSCLE: A Model Update Strategy for Compatible LLM Evolution
Viaarxiv icon

DataComp-LM: In search of the next generation of training sets for language models

Add code
Jun 18, 2024
Figure 1 for DataComp-LM: In search of the next generation of training sets for language models
Figure 2 for DataComp-LM: In search of the next generation of training sets for language models
Figure 3 for DataComp-LM: In search of the next generation of training sets for language models
Figure 4 for DataComp-LM: In search of the next generation of training sets for language models
Viaarxiv icon

CLIP with Quality Captions: A Strong Pretraining for Vision Tasks

Add code
May 14, 2024
Figure 1 for CLIP with Quality Captions: A Strong Pretraining for Vision Tasks
Figure 2 for CLIP with Quality Captions: A Strong Pretraining for Vision Tasks
Figure 3 for CLIP with Quality Captions: A Strong Pretraining for Vision Tasks
Figure 4 for CLIP with Quality Captions: A Strong Pretraining for Vision Tasks
Viaarxiv icon

CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data

Add code
Apr 24, 2024
Figure 1 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Figure 2 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Figure 3 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Figure 4 for CatLIP: CLIP-level Visual Recognition Accuracy with 2.7x Faster Pre-training on Web-scale Image-Text Data
Viaarxiv icon

Weight subcloning: direct initialization of transformers using larger pretrained ones

Add code
Dec 14, 2023
Figure 1 for Weight subcloning: direct initialization of transformers using larger pretrained ones
Figure 2 for Weight subcloning: direct initialization of transformers using larger pretrained ones
Figure 3 for Weight subcloning: direct initialization of transformers using larger pretrained ones
Figure 4 for Weight subcloning: direct initialization of transformers using larger pretrained ones
Viaarxiv icon

Label-efficient Training of Small Task-specific Models by Leveraging Vision Foundation Models

Add code
Nov 30, 2023
Figure 1 for Label-efficient Training of Small Task-specific Models by Leveraging Vision Foundation Models
Figure 2 for Label-efficient Training of Small Task-specific Models by Leveraging Vision Foundation Models
Figure 3 for Label-efficient Training of Small Task-specific Models by Leveraging Vision Foundation Models
Figure 4 for Label-efficient Training of Small Task-specific Models by Leveraging Vision Foundation Models
Viaarxiv icon