Picture for Yejin Bang

Yejin Bang

LLM Internal States Reveal Hallucination Risk Faced With a Query

Add code
Jul 03, 2024
Viaarxiv icon

The Pyramid of Captions

Add code
May 01, 2024
Viaarxiv icon

High-Dimension Human Value Representation in Large Language Models

Add code
Apr 11, 2024
Viaarxiv icon

Measuring Political Bias in Large Language Models: What Is Said and How It Is Said

Add code
Mar 27, 2024
Viaarxiv icon

Mitigating Framing Bias with Polarity Minimization Loss

Add code
Nov 03, 2023
Viaarxiv icon

Survey of Social Bias in Vision-Language Models

Add code
Sep 24, 2023
Viaarxiv icon

Learn What NOT to Learn: Towards Generative Safety in Chatbots

Add code
Apr 25, 2023
Viaarxiv icon

A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity

Add code
Feb 28, 2023
Viaarxiv icon

Casual Conversations v2: Designing a large consent-driven dataset to measure algorithmic bias and robustness

Add code
Nov 10, 2022
Viaarxiv icon

Enabling Classifiers to Make Judgements Explicitly Aligned with Human Values

Add code
Oct 14, 2022
Figure 1 for Enabling Classifiers to Make Judgements Explicitly Aligned with Human Values
Figure 2 for Enabling Classifiers to Make Judgements Explicitly Aligned with Human Values
Figure 3 for Enabling Classifiers to Make Judgements Explicitly Aligned with Human Values
Figure 4 for Enabling Classifiers to Make Judgements Explicitly Aligned with Human Values
Viaarxiv icon