Picture for Akshita Bhagia

Akshita Bhagia

Establishing Task Scaling Laws via Compute-Efficient Model Ladders

Add code
Dec 05, 2024
Viaarxiv icon

OLMoE: Open Mixture-of-Experts Language Models

Add code
Sep 03, 2024
Figure 1 for OLMoE: Open Mixture-of-Experts Language Models
Figure 2 for OLMoE: Open Mixture-of-Experts Language Models
Figure 3 for OLMoE: Open Mixture-of-Experts Language Models
Figure 4 for OLMoE: Open Mixture-of-Experts Language Models
Viaarxiv icon

OLMo: Accelerating the Science of Language Models

Add code
Feb 07, 2024
Figure 1 for OLMo: Accelerating the Science of Language Models
Figure 2 for OLMo: Accelerating the Science of Language Models
Figure 3 for OLMo: Accelerating the Science of Language Models
Figure 4 for OLMo: Accelerating the Science of Language Models
Viaarxiv icon

Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research

Add code
Jan 31, 2024
Figure 1 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 2 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 3 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 4 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Viaarxiv icon

Paloma: A Benchmark for Evaluating Language Model Fit

Add code
Dec 16, 2023
Viaarxiv icon

Catwalk: A Unified Language Model Evaluation Framework for Many Datasets

Add code
Dec 15, 2023
Viaarxiv icon

What's In My Big Data?

Add code
Oct 31, 2023
Figure 1 for What's In My Big Data?
Figure 2 for What's In My Big Data?
Figure 3 for What's In My Big Data?
Figure 4 for What's In My Big Data?
Viaarxiv icon

HINT: Hypernetwork Instruction Tuning for Efficient Zero-Shot Generalisation

Add code
Dec 20, 2022
Viaarxiv icon

Continued Pretraining for Better Zero- and Few-Shot Promptability

Add code
Oct 19, 2022
Figure 1 for Continued Pretraining for Better Zero- and Few-Shot Promptability
Figure 2 for Continued Pretraining for Better Zero- and Few-Shot Promptability
Figure 3 for Continued Pretraining for Better Zero- and Few-Shot Promptability
Figure 4 for Continued Pretraining for Better Zero- and Few-Shot Promptability
Viaarxiv icon

On Advances in Text Generation from Images Beyond Captioning: A Case Study in Self-Rationalization

Add code
May 24, 2022
Figure 1 for On Advances in Text Generation from Images Beyond Captioning: A Case Study in Self-Rationalization
Figure 2 for On Advances in Text Generation from Images Beyond Captioning: A Case Study in Self-Rationalization
Figure 3 for On Advances in Text Generation from Images Beyond Captioning: A Case Study in Self-Rationalization
Figure 4 for On Advances in Text Generation from Images Beyond Captioning: A Case Study in Self-Rationalization
Viaarxiv icon