Picture for Max Bartolo

Max Bartolo

Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models

Add code
Nov 19, 2024
Viaarxiv icon

Understanding Likelihood Over-optimisation in Direct Alignment Algorithms

Add code
Oct 15, 2024
Viaarxiv icon

Improving Reward Models with Synthetic Critiques

Add code
May 31, 2024
Figure 1 for Improving Reward Models with Synthetic Critiques
Figure 2 for Improving Reward Models with Synthetic Critiques
Figure 3 for Improving Reward Models with Synthetic Critiques
Figure 4 for Improving Reward Models with Synthetic Critiques
Viaarxiv icon

Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models

Add code
May 08, 2024
Figure 1 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Figure 2 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Figure 3 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Figure 4 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Viaarxiv icon

The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models

Add code
Apr 24, 2024
Figure 1 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 2 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 3 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 4 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Viaarxiv icon

Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning

Add code
Feb 09, 2024
Figure 1 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 2 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 3 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 4 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Viaarxiv icon

DMLR: Data-centric Machine Learning Research -- Past, Present and Future

Add code
Nov 21, 2023
Figure 1 for DMLR: Data-centric Machine Learning Research -- Past, Present and Future
Figure 2 for DMLR: Data-centric Machine Learning Research -- Past, Present and Future
Figure 3 for DMLR: Data-centric Machine Learning Research -- Past, Present and Future
Viaarxiv icon

Human Feedback is not Gold Standard

Add code
Sep 28, 2023
Figure 1 for Human Feedback is not Gold Standard
Figure 2 for Human Feedback is not Gold Standard
Figure 3 for Human Feedback is not Gold Standard
Figure 4 for Human Feedback is not Gold Standard
Viaarxiv icon

Adversarial Nibbler: A Data-Centric Challenge for Improving the Safety of Text-to-Image Models

Add code
May 22, 2023
Viaarxiv icon

Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality

Add code
Apr 07, 2022
Figure 1 for Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality
Figure 2 for Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality
Figure 3 for Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality
Figure 4 for Winoground: Probing Vision and Language Models for Visio-Linguistic Compositionality
Viaarxiv icon