Picture for Max Bartolo

Max Bartolo

Michael Pokorny

LLMs can implicitly learn from mistakes in-context

Add code
Feb 12, 2025
Viaarxiv icon

Atla Selene Mini: A General Purpose Evaluation Model

Add code
Jan 27, 2025
Viaarxiv icon

Humanity's Last Exam

Add code
Jan 24, 2025
Viaarxiv icon

Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models

Add code
Nov 19, 2024
Figure 1 for Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
Figure 2 for Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
Figure 3 for Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
Figure 4 for Procedural Knowledge in Pretraining Drives Reasoning in Large Language Models
Viaarxiv icon

Understanding Likelihood Over-optimisation in Direct Alignment Algorithms

Add code
Oct 15, 2024
Figure 1 for Understanding Likelihood Over-optimisation in Direct Alignment Algorithms
Figure 2 for Understanding Likelihood Over-optimisation in Direct Alignment Algorithms
Figure 3 for Understanding Likelihood Over-optimisation in Direct Alignment Algorithms
Figure 4 for Understanding Likelihood Over-optimisation in Direct Alignment Algorithms
Viaarxiv icon

Improving Reward Models with Synthetic Critiques

Add code
May 31, 2024
Figure 1 for Improving Reward Models with Synthetic Critiques
Figure 2 for Improving Reward Models with Synthetic Critiques
Figure 3 for Improving Reward Models with Synthetic Critiques
Figure 4 for Improving Reward Models with Synthetic Critiques
Viaarxiv icon

Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models

Add code
May 08, 2024
Figure 1 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Figure 2 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Figure 3 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Figure 4 for Fishing for Magikarp: Automatically Detecting Under-trained Tokens in Large Language Models
Viaarxiv icon

The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models

Add code
Apr 24, 2024
Figure 1 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 2 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 3 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Figure 4 for The PRISM Alignment Project: What Participatory, Representative and Individualised Human Feedback Reveals About the Subjective and Multicultural Alignment of Large Language Models
Viaarxiv icon

Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning

Add code
Feb 09, 2024
Figure 1 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 2 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 3 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Figure 4 for Aya Dataset: An Open-Access Collection for Multilingual Instruction Tuning
Viaarxiv icon

DMLR: Data-centric Machine Learning Research -- Past, Present and Future

Add code
Nov 21, 2023
Figure 1 for DMLR: Data-centric Machine Learning Research -- Past, Present and Future
Figure 2 for DMLR: Data-centric Machine Learning Research -- Past, Present and Future
Figure 3 for DMLR: Data-centric Machine Learning Research -- Past, Present and Future
Viaarxiv icon