Picture for Matthew E. Peters

Matthew E. Peters

OLMo: Accelerating the Science of Language Models

Add code
Feb 07, 2024
Figure 1 for OLMo: Accelerating the Science of Language Models
Figure 2 for OLMo: Accelerating the Science of Language Models
Figure 3 for OLMo: Accelerating the Science of Language Models
Figure 4 for OLMo: Accelerating the Science of Language Models
Viaarxiv icon

Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research

Add code
Jan 31, 2024
Figure 1 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 2 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 3 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Figure 4 for Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research
Viaarxiv icon

ACE: A fast, skillful learned global atmospheric model for climate prediction

Add code
Oct 03, 2023
Figure 1 for ACE: A fast, skillful learned global atmospheric model for climate prediction
Figure 2 for ACE: A fast, skillful learned global atmospheric model for climate prediction
Figure 3 for ACE: A fast, skillful learned global atmospheric model for climate prediction
Figure 4 for ACE: A fast, skillful learned global atmospheric model for climate prediction
Viaarxiv icon

Efficiency Pentathlon: A Standardized Arena for Efficiency Evaluation

Add code
Jul 19, 2023
Viaarxiv icon

Peek Across: Improving Multi-Document Modeling via Cross-Document Question-Answering

Add code
May 24, 2023
Viaarxiv icon

TESS: Text-to-Text Self-Conditioned Simplex Diffusion

Add code
May 15, 2023
Figure 1 for TESS: Text-to-Text Self-Conditioned Simplex Diffusion
Figure 2 for TESS: Text-to-Text Self-Conditioned Simplex Diffusion
Figure 3 for TESS: Text-to-Text Self-Conditioned Simplex Diffusion
Figure 4 for TESS: Text-to-Text Self-Conditioned Simplex Diffusion
Viaarxiv icon

AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models

Add code
Feb 14, 2023
Figure 1 for AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models
Figure 2 for AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models
Figure 3 for AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models
Figure 4 for AdapterSoup: Weight Averaging to Improve Generalization of Pretrained Language Models
Viaarxiv icon

Does Self-Rationalization Improve Robustness to Spurious Correlations?

Add code
Oct 24, 2022
Viaarxiv icon

Attentional Mixtures of Soft Prompt Tuning for Parameter-efficient Multi-task Knowledge Sharing

Add code
May 24, 2022
Figure 1 for Attentional Mixtures of Soft Prompt Tuning for Parameter-efficient Multi-task Knowledge Sharing
Figure 2 for Attentional Mixtures of Soft Prompt Tuning for Parameter-efficient Multi-task Knowledge Sharing
Figure 3 for Attentional Mixtures of Soft Prompt Tuning for Parameter-efficient Multi-task Knowledge Sharing
Figure 4 for Attentional Mixtures of Soft Prompt Tuning for Parameter-efficient Multi-task Knowledge Sharing
Viaarxiv icon

Extracting Latent Steering Vectors from Pretrained Language Models

Add code
May 10, 2022
Figure 1 for Extracting Latent Steering Vectors from Pretrained Language Models
Figure 2 for Extracting Latent Steering Vectors from Pretrained Language Models
Figure 3 for Extracting Latent Steering Vectors from Pretrained Language Models
Figure 4 for Extracting Latent Steering Vectors from Pretrained Language Models
Viaarxiv icon