Picture for Xuefeng Du

Xuefeng Du

Towards Reducible Uncertainty Modeling for Reliable Large Language Model Agents

Add code
Feb 04, 2026
Viaarxiv icon

Harnessing Reasoning Trajectories for Hallucination Detection via Answer-agreement Representation Shaping

Add code
Jan 24, 2026
Viaarxiv icon

Unlocking the Pre-Trained Model as a Dual-Alignment Calibrator for Post-Trained LLMs

Add code
Jan 07, 2026
Viaarxiv icon

Foundations of Unknown-aware Machine Learning

Add code
May 20, 2025
Viaarxiv icon

Safety-Aware Fine-Tuning of Large Language Models

Add code
Oct 13, 2024
Figure 1 for Safety-Aware Fine-Tuning of Large Language Models
Figure 2 for Safety-Aware Fine-Tuning of Large Language Models
Figure 3 for Safety-Aware Fine-Tuning of Large Language Models
Figure 4 for Safety-Aware Fine-Tuning of Large Language Models
Viaarxiv icon

How Reliable Is Human Feedback For Aligning Large Language Models?

Add code
Oct 02, 2024
Figure 1 for How Reliable Is Human Feedback For Aligning Large Language Models?
Figure 2 for How Reliable Is Human Feedback For Aligning Large Language Models?
Figure 3 for How Reliable Is Human Feedback For Aligning Large Language Models?
Figure 4 for How Reliable Is Human Feedback For Aligning Large Language Models?
Viaarxiv icon

HaloScope: Harnessing Unlabeled LLM Generations for Hallucination Detection

Add code
Sep 26, 2024
Viaarxiv icon

Out-of-Distribution Learning with Human Feedback

Add code
Aug 14, 2024
Figure 1 for Out-of-Distribution Learning with Human Feedback
Figure 2 for Out-of-Distribution Learning with Human Feedback
Figure 3 for Out-of-Distribution Learning with Human Feedback
Figure 4 for Out-of-Distribution Learning with Human Feedback
Viaarxiv icon

When and How Does In-Distribution Label Help Out-of-Distribution Detection?

Add code
May 28, 2024
Viaarxiv icon

The Ghanaian NLP Landscape: A First Look

Add code
May 10, 2024
Viaarxiv icon