Picture for Jiacheng Liang

Jiacheng Liang

RASA: Routing-Aware Safety Alignment for Mixture-of-Experts Models

Add code
Feb 04, 2026
Viaarxiv icon

Robustifying Vision-Language Models via Dynamic Token Reweighting

Add code
May 22, 2025
Viaarxiv icon

AutoRAN: Weak-to-Strong Jailbreaking of Large Reasoning Models

Add code
May 16, 2025
Viaarxiv icon

GraphRAG under Fire

Add code
Jan 23, 2025
Viaarxiv icon

WaterPark: A Robustness Assessment of Language Model Watermarking

Add code
Nov 20, 2024
Figure 1 for WaterPark: A Robustness Assessment of Language Model Watermarking
Figure 2 for WaterPark: A Robustness Assessment of Language Model Watermarking
Figure 3 for WaterPark: A Robustness Assessment of Language Model Watermarking
Figure 4 for WaterPark: A Robustness Assessment of Language Model Watermarking
Viaarxiv icon

CopyrightMeter: Revisiting Copyright Protection in Text-to-image Models

Add code
Nov 20, 2024
Figure 1 for CopyrightMeter: Revisiting Copyright Protection in Text-to-image Models
Figure 2 for CopyrightMeter: Revisiting Copyright Protection in Text-to-image Models
Figure 3 for CopyrightMeter: Revisiting Copyright Protection in Text-to-image Models
Figure 4 for CopyrightMeter: Revisiting Copyright Protection in Text-to-image Models
Viaarxiv icon

RobustKV: Defending Large Language Models against Jailbreak Attacks via KV Eviction

Add code
Oct 25, 2024
Figure 1 for RobustKV: Defending Large Language Models against Jailbreak Attacks via KV Eviction
Figure 2 for RobustKV: Defending Large Language Models against Jailbreak Attacks via KV Eviction
Figure 3 for RobustKV: Defending Large Language Models against Jailbreak Attacks via KV Eviction
Figure 4 for RobustKV: Defending Large Language Models against Jailbreak Attacks via KV Eviction
Viaarxiv icon

Buckle Up: Robustifying LLMs at Every Customization Stage via Data Curation

Add code
Oct 03, 2024
Figure 1 for Buckle Up: Robustifying LLMs at Every Customization Stage via Data Curation
Figure 2 for Buckle Up: Robustifying LLMs at Every Customization Stage via Data Curation
Figure 3 for Buckle Up: Robustifying LLMs at Every Customization Stage via Data Curation
Figure 4 for Buckle Up: Robustifying LLMs at Every Customization Stage via Data Curation
Viaarxiv icon

Robustifying Safety-Aligned Large Language Models through Clean Data Curation

Add code
May 31, 2024
Figure 1 for Robustifying Safety-Aligned Large Language Models through Clean Data Curation
Figure 2 for Robustifying Safety-Aligned Large Language Models through Clean Data Curation
Figure 3 for Robustifying Safety-Aligned Large Language Models through Clean Data Curation
Figure 4 for Robustifying Safety-Aligned Large Language Models through Clean Data Curation
Viaarxiv icon

Model Extraction Attacks Revisited

Add code
Dec 08, 2023
Figure 1 for Model Extraction Attacks Revisited
Figure 2 for Model Extraction Attacks Revisited
Figure 3 for Model Extraction Attacks Revisited
Figure 4 for Model Extraction Attacks Revisited
Viaarxiv icon