Picture for Miao Xiong

Miao Xiong

Do LLMs estimate uncertainty well in instruction-following?

Add code
Oct 18, 2024
Viaarxiv icon

FlipAttack: Jailbreak LLMs via Flipping

Add code
Oct 02, 2024
Figure 1 for FlipAttack: Jailbreak LLMs via Flipping
Figure 2 for FlipAttack: Jailbreak LLMs via Flipping
Figure 3 for FlipAttack: Jailbreak LLMs via Flipping
Figure 4 for FlipAttack: Jailbreak LLMs via Flipping
Viaarxiv icon

ID$^3$: Identity-Preserving-yet-Diversified Diffusion Models for Synthetic Face Recognition

Add code
Sep 26, 2024
Figure 1 for ID$^3$: Identity-Preserving-yet-Diversified Diffusion Models for Synthetic Face Recognition
Figure 2 for ID$^3$: Identity-Preserving-yet-Diversified Diffusion Models for Synthetic Face Recognition
Figure 3 for ID$^3$: Identity-Preserving-yet-Diversified Diffusion Models for Synthetic Face Recognition
Figure 4 for ID$^3$: Identity-Preserving-yet-Diversified Diffusion Models for Synthetic Face Recognition
Viaarxiv icon

In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation

Add code
Mar 12, 2024
Figure 1 for In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation
Figure 2 for In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation
Figure 3 for In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation
Figure 4 for In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination Mitigation
Viaarxiv icon

Prompt-and-Align: Prompt-Based Social Alignment for Few-Shot Fake News Detection

Add code
Sep 28, 2023
Viaarxiv icon

Can LLMs Express Their Uncertainty? An Empirical Evaluation of Confidence Elicitation in LLMs

Add code
Jun 22, 2023
Viaarxiv icon

Proximity-Informed Calibration for Deep Neural Networks

Add code
Jun 07, 2023
Viaarxiv icon

GraphCleaner: Detecting Mislabelled Samples in Popular Graph Learning Benchmarks

Add code
May 30, 2023
Viaarxiv icon

Great Models Think Alike: Improving Model Reliability via Inter-Model Latent Agreement

Add code
May 02, 2023
Viaarxiv icon

Trust, but Verify: Using Self-Supervised Probing to Improve Trustworthiness

Add code
Feb 06, 2023
Viaarxiv icon