Picture for Song Mei

Song Mei

GSM-Agent: Understanding Agentic Reasoning Using Controllable Environments

Add code
Sep 26, 2025
Figure 1 for GSM-Agent: Understanding Agentic Reasoning Using Controllable Environments
Figure 2 for GSM-Agent: Understanding Agentic Reasoning Using Controllable Environments
Figure 3 for GSM-Agent: Understanding Agentic Reasoning Using Controllable Environments
Figure 4 for GSM-Agent: Understanding Agentic Reasoning Using Controllable Environments
Viaarxiv icon

Generalization or Hallucination? Understanding Out-of-Context Reasoning in Transformers

Add code
Jun 12, 2025
Viaarxiv icon

SPEED-RL: Faster Training of Reasoning Models via Online Curriculum Learning

Add code
Jun 10, 2025
Viaarxiv icon

OVERT: A Benchmark for Over-Refusal Evaluation on Text-to-Image Models

Add code
May 28, 2025
Viaarxiv icon

Improving LLM Safety Alignment with Dual-Objective Optimization

Add code
Mar 05, 2025
Figure 1 for Improving LLM Safety Alignment with Dual-Objective Optimization
Figure 2 for Improving LLM Safety Alignment with Dual-Objective Optimization
Figure 3 for Improving LLM Safety Alignment with Dual-Objective Optimization
Figure 4 for Improving LLM Safety Alignment with Dual-Objective Optimization
Viaarxiv icon

An Overview of Large Language Models for Statisticians

Add code
Feb 25, 2025
Viaarxiv icon

Implicit Bias of Gradient Descent for Non-Homogeneous Deep Networks

Add code
Feb 22, 2025
Viaarxiv icon

How Do LLMs Perform Two-Hop Reasoning in Context?

Add code
Feb 19, 2025
Viaarxiv icon

A Statistical Theory of Contrastive Pre-training and Multimodal Generative AI

Add code
Jan 08, 2025
Figure 1 for A Statistical Theory of Contrastive Pre-training and Multimodal Generative AI
Figure 2 for A Statistical Theory of Contrastive Pre-training and Multimodal Generative AI
Figure 3 for A Statistical Theory of Contrastive Pre-training and Multimodal Generative AI
Figure 4 for A Statistical Theory of Contrastive Pre-training and Multimodal Generative AI
Viaarxiv icon

Active-Dormant Attention Heads: Mechanistically Demystifying Extreme-Token Phenomena in LLMs

Add code
Oct 17, 2024
Figure 1 for Active-Dormant Attention Heads: Mechanistically Demystifying Extreme-Token Phenomena in LLMs
Figure 2 for Active-Dormant Attention Heads: Mechanistically Demystifying Extreme-Token Phenomena in LLMs
Figure 3 for Active-Dormant Attention Heads: Mechanistically Demystifying Extreme-Token Phenomena in LLMs
Figure 4 for Active-Dormant Attention Heads: Mechanistically Demystifying Extreme-Token Phenomena in LLMs
Viaarxiv icon