Picture for Preslav Nakov

Preslav Nakov

Mohamed bin Zayed University of Artificial Intelligence

When Personalization Tricks Detectors: The Feature-Inversion Trap in Machine-Generated Text Detection

Add code
Oct 14, 2025
Viaarxiv icon

BLAZER: Bootstrapping LLM-based Manipulation Agents with Zero-Shot Data Generation

Add code
Oct 09, 2025
Viaarxiv icon

Toxicity Red-Teaming: Benchmarking LLM Safety in Singapore's Low-Resource Languages

Add code
Sep 18, 2025
Figure 1 for Toxicity Red-Teaming: Benchmarking LLM Safety in Singapore's Low-Resource Languages
Figure 2 for Toxicity Red-Teaming: Benchmarking LLM Safety in Singapore's Low-Resource Languages
Figure 3 for Toxicity Red-Teaming: Benchmarking LLM Safety in Singapore's Low-Resource Languages
Figure 4 for Toxicity Red-Teaming: Benchmarking LLM Safety in Singapore's Low-Resource Languages
Viaarxiv icon

Explicit and Implicit Data Augmentation for Social Event Detection

Add code
Sep 04, 2025
Viaarxiv icon

BALSAM: A Platform for Benchmarking Arabic Large Language Models

Add code
Jul 30, 2025
Viaarxiv icon

FRaN-X: FRaming and Narratives-eXplorer

Add code
Jul 09, 2025
Viaarxiv icon

Profiling News Media for Factuality and Bias Using LLMs and the Fact-Checking Methodology of Human Experts

Add code
Jun 14, 2025
Viaarxiv icon

Atomic Reasoning for Scientific Table Claim Verification

Add code
Jun 08, 2025
Viaarxiv icon

Faithfulness-Aware Uncertainty Quantification for Fact-Checking the Output of Retrieval Augmented Generation

Add code
May 28, 2025
Viaarxiv icon

VSCBench: Bridging the Gap in Vision-Language Model Safety Calibration

Add code
May 26, 2025
Viaarxiv icon