Picture for Alex Mei

Alex Mei

ASSERT: Automated Safety Scenario Red Teaming for Evaluating the Robustness of Large Language Models

Add code
Oct 14, 2023
Viaarxiv icon

Let's Think Frame by Frame: Evaluating Video Chain of Thought with Video Infilling and Prediction

Add code
May 23, 2023
Viaarxiv icon

Visual Chain of Thought: Bridging Logical Gaps with Multimodal Infillings

Add code
May 03, 2023
Viaarxiv icon

Users are the North Star for AI Transparency

Add code
Mar 09, 2023
Viaarxiv icon

Foveate, Attribute, and Rationalize: Towards Safe and Trustworthy AI

Add code
Dec 19, 2022
Viaarxiv icon

Mitigating Covertly Unsafe Text within Natural Language Systems

Add code
Oct 17, 2022
Figure 1 for Mitigating Covertly Unsafe Text within Natural Language Systems
Figure 2 for Mitigating Covertly Unsafe Text within Natural Language Systems
Figure 3 for Mitigating Covertly Unsafe Text within Natural Language Systems
Figure 4 for Mitigating Covertly Unsafe Text within Natural Language Systems
Viaarxiv icon