Picture for Giulio Zizzo

Giulio Zizzo

Blue Teaming Function-Calling Agents

Add code
Jan 14, 2026
Viaarxiv icon

Step-Tagging: Toward controlling the generation of Language Reasoning Models through step monitoring

Add code
Dec 16, 2025
Viaarxiv icon

Pre-Hoc Predictions in AutoML: Leveraging LLMs to Enhance Model Selection and Benchmarking for Tabular datasets

Add code
Oct 02, 2025
Viaarxiv icon

Activated LoRA: Fine-tuned LLMs for Intrinsics

Add code
Apr 16, 2025
Viaarxiv icon

MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming

Add code
Mar 08, 2025
Figure 1 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Figure 2 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Figure 3 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Figure 4 for MAD-MAX: Modular And Diverse Malicious Attack MiXtures for Automated LLM Red Teaming
Viaarxiv icon

Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs

Add code
Feb 21, 2025
Figure 1 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Figure 2 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Figure 3 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Figure 4 for Adversarial Prompt Evaluation: Systematic Benchmarking of Guardrails Against Prompt Input Attacks on LLMs
Viaarxiv icon

Granite Guardian

Add code
Dec 10, 2024
Figure 1 for Granite Guardian
Figure 2 for Granite Guardian
Figure 3 for Granite Guardian
Figure 4 for Granite Guardian
Viaarxiv icon

HarmLevelBench: Evaluating Harm-Level Compliance and the Impact of Quantization on Model Alignment

Add code
Nov 11, 2024
Viaarxiv icon

Assessing the Impact of Packing on Machine Learning-Based Malware Detection and Classification Systems

Add code
Oct 31, 2024
Figure 1 for Assessing the Impact of Packing on Machine Learning-Based Malware Detection and Classification Systems
Figure 2 for Assessing the Impact of Packing on Machine Learning-Based Malware Detection and Classification Systems
Figure 3 for Assessing the Impact of Packing on Machine Learning-Based Malware Detection and Classification Systems
Figure 4 for Assessing the Impact of Packing on Machine Learning-Based Malware Detection and Classification Systems
Viaarxiv icon

Towards Assurance of LLM Adversarial Robustness using Ontology-Driven Argumentation

Add code
Oct 10, 2024
Figure 1 for Towards Assurance of LLM Adversarial Robustness using Ontology-Driven Argumentation
Figure 2 for Towards Assurance of LLM Adversarial Robustness using Ontology-Driven Argumentation
Figure 3 for Towards Assurance of LLM Adversarial Robustness using Ontology-Driven Argumentation
Figure 4 for Towards Assurance of LLM Adversarial Robustness using Ontology-Driven Argumentation
Viaarxiv icon