Picture for Kuofeng Gao

Kuofeng Gao

Embedding Self-Correction as an Inherent Ability in Large Language Models for Enhanced Mathematical Reasoning

Add code
Oct 14, 2024
Figure 1 for Embedding Self-Correction as an Inherent Ability in Large Language Models for Enhanced Mathematical Reasoning
Figure 2 for Embedding Self-Correction as an Inherent Ability in Large Language Models for Enhanced Mathematical Reasoning
Figure 3 for Embedding Self-Correction as an Inherent Ability in Large Language Models for Enhanced Mathematical Reasoning
Figure 4 for Embedding Self-Correction as an Inherent Ability in Large Language Models for Enhanced Mathematical Reasoning
Viaarxiv icon

Denial-of-Service Poisoning Attacks against Large Language Models

Add code
Oct 14, 2024
Figure 1 for Denial-of-Service Poisoning Attacks against Large Language Models
Figure 2 for Denial-of-Service Poisoning Attacks against Large Language Models
Figure 3 for Denial-of-Service Poisoning Attacks against Large Language Models
Figure 4 for Denial-of-Service Poisoning Attacks against Large Language Models
Viaarxiv icon

PointNCBW: Towards Dataset Ownership Verification for Point Clouds via Negative Clean-label Backdoor Watermark

Add code
Aug 10, 2024
Viaarxiv icon

Video Watermarking: Safeguarding Your Video from (Unauthorized) Annotations by Video-based LLMs

Add code
Jul 02, 2024
Viaarxiv icon

Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas

Add code
Jun 08, 2024
Figure 1 for Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas
Figure 2 for Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas
Figure 3 for Deconstructing The Ethics of Large Language Models from Long-standing Issues to New-emerging Dilemmas
Viaarxiv icon

Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision Transformers

Add code
May 17, 2024
Viaarxiv icon

Adversarial Robustness for Visual Grounding of Multimodal Large Language Models

Add code
May 16, 2024
Viaarxiv icon

Energy-Latency Manipulation of Multi-modal Large Language Models via Verbose Samples

Add code
Apr 25, 2024
Viaarxiv icon

FMM-Attack: A Flow-based Multi-modal Adversarial Attack on Video-based LLMs

Add code
Mar 21, 2024
Viaarxiv icon

Inducing High Energy-Latency of Large Vision-Language Models with Verbose Images

Add code
Jan 20, 2024
Viaarxiv icon