Picture for Gaotang Li

Gaotang Li

Do VLMs Have a Moral Backbone? A Study on the Fragile Morality of Vision-Language Models

Add code
Jan 23, 2026
Viaarxiv icon

Agentic Reasoning for Large Language Models

Add code
Jan 18, 2026
Viaarxiv icon

ALERT: Zero-shot LLM Jailbreak Detection via Internal Discrepancy Amplification

Add code
Jan 07, 2026
Viaarxiv icon

Stabilizing Reinforcement Learning for Honesty Alignment in Language Models on Deductive Reasoning

Add code
Nov 12, 2025
Viaarxiv icon

Beyond Log Likelihood: Probability-Based Objectives for Supervised Fine-Tuning across the Model Capability Continuum

Add code
Oct 01, 2025
Viaarxiv icon

Graph Homophily Booster: Rethinking the Role of Discrete Features on Heterophilic Graphs

Add code
Sep 16, 2025
Viaarxiv icon

Saffron-1: Towards an Inference Scaling Paradigm for LLM Safety Assurance

Add code
Jun 06, 2025
Viaarxiv icon

MORALISE: A Structured Benchmark for Moral Alignment in Visual Language Models

Add code
May 20, 2025
Figure 1 for MORALISE: A Structured Benchmark for Moral Alignment in Visual Language Models
Figure 2 for MORALISE: A Structured Benchmark for Moral Alignment in Visual Language Models
Figure 3 for MORALISE: A Structured Benchmark for Moral Alignment in Visual Language Models
Figure 4 for MORALISE: A Structured Benchmark for Moral Alignment in Visual Language Models
Viaarxiv icon

RM-R1: Reward Modeling as Reasoning

Add code
May 05, 2025
Viaarxiv icon

Taming Knowledge Conflicts in Language Models

Add code
Mar 14, 2025
Figure 1 for Taming Knowledge Conflicts in Language Models
Figure 2 for Taming Knowledge Conflicts in Language Models
Figure 3 for Taming Knowledge Conflicts in Language Models
Figure 4 for Taming Knowledge Conflicts in Language Models
Viaarxiv icon