Picture for Percy Liang

Percy Liang

Shammie

Reliable and Efficient Amortized Model-based Evaluation

Add code
Mar 17, 2025
Viaarxiv icon

Fine-Tuning Vision-Language-Action Models: Optimizing Speed and Success

Add code
Feb 27, 2025
Viaarxiv icon

The Mighty ToRR: A Benchmark for Table Reasoning and Robustness

Add code
Feb 26, 2025
Viaarxiv icon

Independence Tests for Language Models

Add code
Feb 17, 2025
Viaarxiv icon

Auditing Prompt Caching in Language Model APIs

Add code
Feb 11, 2025
Viaarxiv icon

Eliciting Language Model Behaviors with Investigator Agents

Add code
Feb 03, 2025
Viaarxiv icon

Language Models Prefer What They Know: Relative Confidence Estimation via Confidence Preferences

Add code
Feb 03, 2025
Viaarxiv icon

s1: Simple test-time scaling

Add code
Jan 31, 2025
Figure 1 for s1: Simple test-time scaling
Figure 2 for s1: Simple test-time scaling
Figure 3 for s1: Simple test-time scaling
Figure 4 for s1: Simple test-time scaling
Viaarxiv icon

International AI Safety Report

Add code
Jan 29, 2025
Viaarxiv icon

Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice

Add code
Dec 09, 2024
Figure 1 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 2 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 3 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Figure 4 for Machine Unlearning Doesn't Do What You Think: Lessons for Generative AI Policy, Research, and Practice
Viaarxiv icon