Picture for Yanyan Zhao

Yanyan Zhao

Chain of Strategy Optimization Makes Large Language Models Better Emotional Supporter

Add code
Mar 07, 2025
Viaarxiv icon

Beware of Your Po! Measuring and Mitigating AI Safety Risks in Role-Play Fine-Tuning of LLMs

Add code
Feb 28, 2025
Viaarxiv icon

Can Large Language Models Understand You Better? An MBTI Personality Detection Dataset Aligned with Population Traits

Add code
Dec 17, 2024
Viaarxiv icon

Separate the Wheat from the Chaff: A Post-Hoc Approach to Safety Re-Alignment for Fine-Tuned Language Models

Add code
Dec 15, 2024
Figure 1 for Separate the Wheat from the Chaff: A Post-Hoc Approach to Safety Re-Alignment for Fine-Tuned Language Models
Figure 2 for Separate the Wheat from the Chaff: A Post-Hoc Approach to Safety Re-Alignment for Fine-Tuned Language Models
Figure 3 for Separate the Wheat from the Chaff: A Post-Hoc Approach to Safety Re-Alignment for Fine-Tuned Language Models
Figure 4 for Separate the Wheat from the Chaff: A Post-Hoc Approach to Safety Re-Alignment for Fine-Tuned Language Models
Viaarxiv icon

Data Uncertainty-Aware Learning for Multimodal Aspect-based Sentiment Analysis

Add code
Dec 02, 2024
Viaarxiv icon

Lens: Rethinking Multilingual Enhancement for Large Language Models

Add code
Oct 06, 2024
Figure 1 for Lens: Rethinking Multilingual Enhancement for Large Language Models
Figure 2 for Lens: Rethinking Multilingual Enhancement for Large Language Models
Figure 3 for Lens: Rethinking Multilingual Enhancement for Large Language Models
Figure 4 for Lens: Rethinking Multilingual Enhancement for Large Language Models
Viaarxiv icon

Large Language Models Meet Text-Centric Multimodal Sentiment Analysis: A Survey

Add code
Jun 12, 2024
Viaarxiv icon

RKLD: Reverse KL-Divergence-based Knowledge Distillation for Unlearning Personal Information in Large Language Models

Add code
Jun 04, 2024
Viaarxiv icon

Towards Comprehensive and Efficient Post Safety Alignment of Large Language Models via Safety Patching

Add code
May 22, 2024
Figure 1 for Towards Comprehensive and Efficient Post Safety Alignment of Large Language Models via Safety Patching
Figure 2 for Towards Comprehensive and Efficient Post Safety Alignment of Large Language Models via Safety Patching
Figure 3 for Towards Comprehensive and Efficient Post Safety Alignment of Large Language Models via Safety Patching
Figure 4 for Towards Comprehensive and Efficient Post Safety Alignment of Large Language Models via Safety Patching
Viaarxiv icon

Vanilla Transformers are Transfer Capability Teachers

Add code
Mar 04, 2024
Viaarxiv icon