Picture for Yuning Mao

Yuning Mao

Jack

The Llama 4 Herd: Architecture, Training, Evaluation, and Deployment Notes

Add code
Jan 15, 2026
Viaarxiv icon

Dr. Zero: Self-Evolving Search Agents without Training Data

Add code
Jan 11, 2026
Viaarxiv icon

High Accuracy, Less Talk (HALT): Reliable LLMs through Capability-Aligned Finetuning

Add code
Jun 04, 2025
Viaarxiv icon

Diversity-driven Data Selection for Language Model Tuning through Sparse Autoencoder

Add code
Feb 19, 2025
Viaarxiv icon

Improving Model Factuality with Fine-grained Critique-based Evaluator

Add code
Oct 24, 2024
Viaarxiv icon

The Llama 3 Herd of Models

Add code
Jul 31, 2024
Viaarxiv icon

Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts

Add code
Feb 26, 2024
Figure 1 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Figure 2 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Figure 3 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Figure 4 for Rainbow Teaming: Open-Ended Generation of Diverse Adversarial Prompts
Viaarxiv icon

RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training

Add code
Dec 07, 2023
Figure 1 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 2 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 3 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Figure 4 for RoAST: Robustifying Language Models via Adversarial Perturbation with Selective Training
Viaarxiv icon

Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations

Add code
Dec 07, 2023
Figure 1 for Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations
Figure 2 for Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations
Figure 3 for Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations
Figure 4 for Llama Guard: LLM-based Input-Output Safeguard for Human-AI Conversations
Viaarxiv icon

MART: Improving LLM Safety with Multi-round Automatic Red-Teaming

Add code
Nov 13, 2023
Figure 1 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 2 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 3 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Figure 4 for MART: Improving LLM Safety with Multi-round Automatic Red-Teaming
Viaarxiv icon