Picture for Andres Algaba

Andres Algaba

Michael Pokorny

Probing the Trajectories of Reasoning Traces in Large Language Models

Add code
Jan 30, 2026
Viaarxiv icon

Structurally Human, Semantically Biased: Detecting LLM-Generated References with Embeddings and GNNs

Add code
Jan 28, 2026
Viaarxiv icon

Benchmarks Saturate When The Model Gets Smarter Than The Judge

Add code
Jan 27, 2026
Viaarxiv icon

Estimating problem difficulty without ground truth using Large Language Model comparisons

Add code
Dec 16, 2025
Viaarxiv icon

Decision-centric fairness: Evaluation and optimization for resource allocation problems

Add code
Apr 29, 2025
Figure 1 for Decision-centric fairness: Evaluation and optimization for resource allocation problems
Figure 2 for Decision-centric fairness: Evaluation and optimization for resource allocation problems
Figure 3 for Decision-centric fairness: Evaluation and optimization for resource allocation problems
Figure 4 for Decision-centric fairness: Evaluation and optimization for resource allocation problems
Viaarxiv icon

How Deep Do Large Language Models Internalize Scientific Literature and Citation Practices?

Add code
Apr 03, 2025
Figure 1 for How Deep Do Large Language Models Internalize Scientific Literature and Citation Practices?
Figure 2 for How Deep Do Large Language Models Internalize Scientific Literature and Citation Practices?
Figure 3 for How Deep Do Large Language Models Internalize Scientific Literature and Citation Practices?
Figure 4 for How Deep Do Large Language Models Internalize Scientific Literature and Citation Practices?
Viaarxiv icon

Flexible Counterfactual Explanations with Generative Models

Add code
Feb 24, 2025
Figure 1 for Flexible Counterfactual Explanations with Generative Models
Figure 2 for Flexible Counterfactual Explanations with Generative Models
Figure 3 for Flexible Counterfactual Explanations with Generative Models
Figure 4 for Flexible Counterfactual Explanations with Generative Models
Viaarxiv icon

The Relationship Between Reasoning and Performance in Large Language Models -- o3 (mini) Thinks Harder, Not Longer

Add code
Feb 21, 2025
Viaarxiv icon

Humanity's Last Exam

Add code
Jan 24, 2025
Viaarxiv icon

Large Language Models Reflect Human Citation Patterns with a Heightened Citation Bias

Add code
May 29, 2024
Viaarxiv icon