Picture for Micah Goldblum

Micah Goldblum

Privacy-Preserving Mechanisms Enable Cheap Verifiable Inference of LLMs

Add code
Feb 19, 2026
Viaarxiv icon

Dynamic Delayed Tree Expansion For Improved Multi-Path Speculative Decoding

Add code
Feb 19, 2026
Viaarxiv icon

Multi-Token Prediction via Self-Distillation

Add code
Feb 05, 2026
Viaarxiv icon

Closing the Train-Test Gap in World Models for Gradient-Based Planning

Add code
Dec 10, 2025
Figure 1 for Closing the Train-Test Gap in World Models for Gradient-Based Planning
Figure 2 for Closing the Train-Test Gap in World Models for Gradient-Based Planning
Figure 3 for Closing the Train-Test Gap in World Models for Gradient-Based Planning
Figure 4 for Closing the Train-Test Gap in World Models for Gradient-Based Planning
Viaarxiv icon

Incoherent Beliefs & Inconsistent Actions in Large Language Models

Add code
Nov 17, 2025
Viaarxiv icon

Preserving Cross-Modal Consistency for CLIP-based Class-Incremental Learning

Add code
Nov 14, 2025
Figure 1 for Preserving Cross-Modal Consistency for CLIP-based Class-Incremental Learning
Figure 2 for Preserving Cross-Modal Consistency for CLIP-based Class-Incremental Learning
Figure 3 for Preserving Cross-Modal Consistency for CLIP-based Class-Incremental Learning
Figure 4 for Preserving Cross-Modal Consistency for CLIP-based Class-Incremental Learning
Viaarxiv icon

Teaching Pretrained Language Models to Think Deeper with Retrofitted Recurrence

Add code
Nov 10, 2025
Viaarxiv icon

Far from the Shallow: Brain-Predictive Reasoning Embedding through Residual Disentanglement

Add code
Oct 26, 2025
Viaarxiv icon

Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning

Add code
Jul 22, 2025
Figure 1 for Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning
Figure 2 for Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning
Figure 3 for Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning
Figure 4 for Zebra-CoT: A Dataset for Interleaved Vision Language Reasoning
Viaarxiv icon

Small Batch Size Training for Language Models: When Vanilla SGD Works, and Why Gradient Accumulation Is Wasteful

Add code
Jul 09, 2025
Viaarxiv icon