Picture for David Lindner

David Lindner

Aligned, Orthogonal or In-conflict: When can we safely optimize Chain-of-Thought?

Add code
Mar 31, 2026
Viaarxiv icon

Quantifying the Necessity of Chain of Thought through Opaque Serial Depth

Add code
Mar 10, 2026
Viaarxiv icon

Evaluating and Understanding Scheming Propensity in LLM Agents

Add code
Mar 02, 2026
Viaarxiv icon

Frontier Models Can Take Actions at Low Probabilities

Add code
Mar 02, 2026
Viaarxiv icon

Stress-Testing Alignment Audits With Prompt-Level Strategic Deception

Add code
Feb 09, 2026
Viaarxiv icon

Practical challenges of control monitoring in frontier AI deployments

Add code
Dec 15, 2025
Viaarxiv icon

Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety

Add code
Jul 15, 2025
Figure 1 for Chain of Thought Monitorability: A New and Fragile Opportunity for AI Safety
Viaarxiv icon

Early Signs of Steganographic Capabilities in Frontier LLMs

Add code
Jul 03, 2025
Figure 1 for Early Signs of Steganographic Capabilities in Frontier LLMs
Figure 2 for Early Signs of Steganographic Capabilities in Frontier LLMs
Figure 3 for Early Signs of Steganographic Capabilities in Frontier LLMs
Figure 4 for Early Signs of Steganographic Capabilities in Frontier LLMs
Viaarxiv icon

Evaluating Frontier Models for Stealth and Situational Awareness

Add code
May 02, 2025
Viaarxiv icon

An Approach to Technical AGI Safety and Security

Add code
Apr 02, 2025
Viaarxiv icon