Picture for Alexander Wan

Alexander Wan

What Evidence Do Language Models Find Convincing?

Add code
Feb 19, 2024
Viaarxiv icon

Poisoning Language Models During Instruction Tuning

Add code
May 01, 2023
Viaarxiv icon

GLUECons: A Generic Benchmark for Learning Under Constraints

Add code
Feb 16, 2023
Viaarxiv icon