Picture for Hongbang Yuan

Hongbang Yuan

RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment

Add code
Dec 18, 2024
Figure 1 for RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment
Figure 2 for RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment
Figure 3 for RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment
Figure 4 for RAG-RewardBench: Benchmarking Reward Models in Retrieval Augmented Generation for Preference Alignment
Viaarxiv icon

Towards Robust Knowledge Unlearning: An Adversarial Framework for Assessing and Improving Unlearning Robustness in Large Language Models

Add code
Aug 20, 2024
Viaarxiv icon

Beyond Under-Alignment: Atomic Preference Enhanced Factuality Tuning for Large Language Models

Add code
Jun 18, 2024
Viaarxiv icon

RWKU: Benchmarking Real-World Knowledge Unlearning for Large Language Models

Add code
Jun 16, 2024
Viaarxiv icon

Whispers that Shake Foundations: Analyzing and Mitigating False Premise Hallucinations in Large Language Models

Add code
Feb 29, 2024
Viaarxiv icon

Cutting Off the Head Ends the Conflict: A Mechanism for Interpreting and Mitigating Knowledge Conflicts in Language Models

Add code
Feb 28, 2024
Viaarxiv icon