Picture for Jakob N. Foerster

Jakob N. Foerster

University of Oxford

Beyond the Boundaries of Proximal Policy Optimization

Add code
Nov 01, 2024
Figure 1 for Beyond the Boundaries of Proximal Policy Optimization
Figure 2 for Beyond the Boundaries of Proximal Policy Optimization
Figure 3 for Beyond the Boundaries of Proximal Policy Optimization
Figure 4 for Beyond the Boundaries of Proximal Policy Optimization
Viaarxiv icon

Opponent Shaping for Antibody Development

Add code
Sep 19, 2024
Viaarxiv icon

Discovering Minimal Reinforcement Learning Environments

Add code
Jun 18, 2024
Viaarxiv icon

HelloFresh: LLM Evaluations on Streams of Real-World Human Editorial Actions across X Community Notes and Wikipedia edits

Add code
Jun 05, 2024
Figure 1 for HelloFresh: LLM Evaluations on Streams of Real-World Human Editorial Actions across X Community Notes and Wikipedia edits
Figure 2 for HelloFresh: LLM Evaluations on Streams of Real-World Human Editorial Actions across X Community Notes and Wikipedia edits
Figure 3 for HelloFresh: LLM Evaluations on Streams of Real-World Human Editorial Actions across X Community Notes and Wikipedia edits
Figure 4 for HelloFresh: LLM Evaluations on Streams of Real-World Human Editorial Actions across X Community Notes and Wikipedia edits
Viaarxiv icon

SMACv2: An Improved Benchmark for Cooperative Multi-Agent Reinforcement Learning

Add code
Dec 14, 2022
Figure 1 for SMACv2: An Improved Benchmark for Cooperative Multi-Agent Reinforcement Learning
Figure 2 for SMACv2: An Improved Benchmark for Cooperative Multi-Agent Reinforcement Learning
Figure 3 for SMACv2: An Improved Benchmark for Cooperative Multi-Agent Reinforcement Learning
Figure 4 for SMACv2: An Improved Benchmark for Cooperative Multi-Agent Reinforcement Learning
Viaarxiv icon

Game-Theoretical Perspectives on Active Equilibria: A Preferred Solution Concept over Nash Equilibria

Add code
Oct 28, 2022
Viaarxiv icon

Learning to Optimize Quasi-Newton Methods

Add code
Oct 11, 2022
Figure 1 for Learning to Optimize Quasi-Newton Methods
Figure 2 for Learning to Optimize Quasi-Newton Methods
Figure 3 for Learning to Optimize Quasi-Newton Methods
Figure 4 for Learning to Optimize Quasi-Newton Methods
Viaarxiv icon

Illusionary Attacks on Sequential Decision Makers and Countermeasures

Add code
Jul 20, 2022
Figure 1 for Illusionary Attacks on Sequential Decision Makers and Countermeasures
Figure 2 for Illusionary Attacks on Sequential Decision Makers and Countermeasures
Figure 3 for Illusionary Attacks on Sequential Decision Makers and Countermeasures
Figure 4 for Illusionary Attacks on Sequential Decision Makers and Countermeasures
Viaarxiv icon

K-level Reasoning for Zero-Shot Coordination in Hanabi

Add code
Jul 14, 2022
Figure 1 for K-level Reasoning for Zero-Shot Coordination in Hanabi
Figure 2 for K-level Reasoning for Zero-Shot Coordination in Hanabi
Figure 3 for K-level Reasoning for Zero-Shot Coordination in Hanabi
Figure 4 for K-level Reasoning for Zero-Shot Coordination in Hanabi
Viaarxiv icon

Influencing Long-Term Behavior in Multiagent Reinforcement Learning

Add code
Mar 07, 2022
Figure 1 for Influencing Long-Term Behavior in Multiagent Reinforcement Learning
Figure 2 for Influencing Long-Term Behavior in Multiagent Reinforcement Learning
Figure 3 for Influencing Long-Term Behavior in Multiagent Reinforcement Learning
Figure 4 for Influencing Long-Term Behavior in Multiagent Reinforcement Learning
Viaarxiv icon