Picture for Daoyuan Wu

Daoyuan Wu

GuidedBench: Equipping Jailbreak Evaluation with Guidelines

Add code
Feb 24, 2025
Viaarxiv icon

DeFiScope: Detecting Various DeFi Price Manipulations with LLM Reasoning

Add code
Feb 17, 2025
Viaarxiv icon

API-guided Dataset Synthesis to Finetune Large Code Models

Add code
Aug 15, 2024
Figure 1 for API-guided Dataset Synthesis to Finetune Large Code Models
Figure 2 for API-guided Dataset Synthesis to Finetune Large Code Models
Figure 3 for API-guided Dataset Synthesis to Finetune Large Code Models
Figure 4 for API-guided Dataset Synthesis to Finetune Large Code Models
Viaarxiv icon

SelfDefend: LLMs Can Defend Themselves against Jailbreaking in a Practical Manner

Add code
Jun 08, 2024
Viaarxiv icon

PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation

Add code
May 04, 2024
Figure 1 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Figure 2 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Figure 3 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Figure 4 for PropertyGPT: LLM-driven Formal Verification of Smart Contracts through Retrieval-Augmented Property Generation
Viaarxiv icon

Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs

Add code
Apr 27, 2024
Figure 1 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Figure 2 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Figure 3 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Figure 4 for Testing and Understanding Erroneous Planning in LLM Agents through Synthesized User Inputs
Viaarxiv icon

LLMs Can Defend Themselves Against Jailbreaking in a Practical Manner: A Vision Paper

Add code
Mar 04, 2024
Viaarxiv icon

LLM4Vuln: A Unified Evaluation Framework for Decoupling and Enhancing LLMs' Vulnerability Reasoning

Add code
Jan 29, 2024
Figure 1 for LLM4Vuln: A Unified Evaluation Framework for Decoupling and Enhancing LLMs' Vulnerability Reasoning
Figure 2 for LLM4Vuln: A Unified Evaluation Framework for Decoupling and Enhancing LLMs' Vulnerability Reasoning
Figure 3 for LLM4Vuln: A Unified Evaluation Framework for Decoupling and Enhancing LLMs' Vulnerability Reasoning
Figure 4 for LLM4Vuln: A Unified Evaluation Framework for Decoupling and Enhancing LLMs' Vulnerability Reasoning
Viaarxiv icon

VRPTEST: Evaluating Visual Referring Prompting in Large Multimodal Models

Add code
Dec 07, 2023
Figure 1 for VRPTEST: Evaluating Visual Referring Prompting in Large Multimodal Models
Figure 2 for VRPTEST: Evaluating Visual Referring Prompting in Large Multimodal Models
Figure 3 for VRPTEST: Evaluating Visual Referring Prompting in Large Multimodal Models
Figure 4 for VRPTEST: Evaluating Visual Referring Prompting in Large Multimodal Models
Viaarxiv icon

Split and Merge: Aligning Position Biases in Large Language Model based Evaluators

Add code
Oct 09, 2023
Figure 1 for Split and Merge: Aligning Position Biases in Large Language Model based Evaluators
Figure 2 for Split and Merge: Aligning Position Biases in Large Language Model based Evaluators
Figure 3 for Split and Merge: Aligning Position Biases in Large Language Model based Evaluators
Figure 4 for Split and Merge: Aligning Position Biases in Large Language Model based Evaluators
Viaarxiv icon