Picture for Neil Zhenqiang Gong

Neil Zhenqiang Gong

Making LLMs Vulnerable to Prompt Injection via Poisoning Alignment

Add code
Oct 18, 2024
Viaarxiv icon

Automatically Generating Visual Hallucination Test Cases for Multimodal Large Language Models

Add code
Oct 15, 2024
Viaarxiv icon

StringLLM: Understanding the String Processing Capability of Large Language Models

Add code
Oct 02, 2024
Viaarxiv icon

A General Framework for Data-Use Auditing of ML Models

Add code
Jul 21, 2024
Viaarxiv icon

Refusing Safe Prompts for Multi-modal Large Language Models

Add code
Jul 12, 2024
Viaarxiv icon

Tracing Back the Malicious Clients in Poisoning Attacks to Federated Learning

Add code
Jul 09, 2024
Viaarxiv icon

Certifiably Robust Image Watermark

Add code
Jul 04, 2024
Viaarxiv icon

Self-Cognition in Large Language Models: An Exploratory Study

Add code
Jul 01, 2024
Figure 1 for Self-Cognition in Large Language Models: An Exploratory Study
Figure 2 for Self-Cognition in Large Language Models: An Exploratory Study
Figure 3 for Self-Cognition in Large Language Models: An Exploratory Study
Figure 4 for Self-Cognition in Large Language Models: An Exploratory Study
Viaarxiv icon

ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods

Add code
Jun 23, 2024
Figure 1 for ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Figure 2 for ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Figure 3 for ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Figure 4 for ReCaLL: Membership Inference via Relative Conditional Log-Likelihoods
Viaarxiv icon

AudioMarkBench: Benchmarking Robustness of Audio Watermarking

Add code
Jun 11, 2024
Viaarxiv icon