Picture for Xiaoling Wang

Xiaoling Wang

Model-Free Output Feedback Stabilization via Policy Gradient Methods

Add code
Jan 29, 2026
Viaarxiv icon

Output Feedback Stabilization of Linear Systems via Policy Gradient Methods

Add code
Jan 27, 2026
Viaarxiv icon

Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education

Add code
Nov 18, 2025
Figure 1 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Figure 2 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Figure 3 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Figure 4 for Unified Defense for Large Language Models against Jailbreak and Fine-Tuning Attacks in Education
Viaarxiv icon

Hierarchical Safety Realignment: Lightweight Restoration of Safety in Pruned Large Vision-Language Models

Add code
May 22, 2025
Viaarxiv icon

AutoMedEval: Harnessing Language Models for Automatic Medical Capability Evaluation

Add code
May 17, 2025
Viaarxiv icon

Unified Attacks to Large Language Model Watermarks: Spoofing and Scrubbing in Unauthorized Knowledge Distillation

Add code
Apr 24, 2025
Viaarxiv icon

HSACNet: Hierarchical Scale-Aware Consistency Regularized Semi-Supervised Change Detection

Add code
Apr 18, 2025
Viaarxiv icon

Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks

Add code
Jan 18, 2025
Figure 1 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Figure 2 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Figure 3 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Figure 4 for Latent-space adversarial training with post-aware calibration for defending large language models against jailbreak attacks
Viaarxiv icon

Hierarchical Divide-and-Conquer for Fine-Grained Alignment in LLM-Based Medical Evaluation

Add code
Jan 12, 2025
Viaarxiv icon

NLSR: Neuron-Level Safety Realignment of Large Language Models Against Harmful Fine-Tuning

Add code
Dec 17, 2024
Viaarxiv icon