Picture for Ari Holtzman

Ari Holtzman

Byte Latent Transformer: Patches Scale Better Than Tokens

Add code
Dec 13, 2024
Viaarxiv icon

Forking Paths in Neural Text Generation

Add code
Dec 10, 2024
Viaarxiv icon

Benchmarks as Microscopes: A Call for Model Metrology

Add code
Jul 22, 2024
Viaarxiv icon

MUSE: Machine Unlearning Six-Way Evaluation for Language Models

Add code
Jul 08, 2024
Viaarxiv icon

Predicting vs. Acting: A Trade-off Between World Modeling & Agent Modeling

Add code
Jul 02, 2024
Viaarxiv icon

Nearest Neighbor Speculative Decoding for LLM Generation and Attribution

Add code
May 29, 2024
Figure 1 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 2 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 3 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Figure 4 for Nearest Neighbor Speculative Decoding for LLM Generation and Attribution
Viaarxiv icon

Superposed Decoding: Multiple Generations from a Single Autoregressive Inference Pass

Add code
May 29, 2024
Viaarxiv icon

CacheGen: Fast Context Loading for Language Model Applications

Add code
Oct 11, 2023
Viaarxiv icon

How FaR Are Large Language Models From Agents with Theory-of-Mind?

Add code
Oct 04, 2023
Figure 1 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 2 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 3 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Figure 4 for How FaR Are Large Language Models From Agents with Theory-of-Mind?
Viaarxiv icon

Generative Models as a Complex Systems Science: How can we make sense of large language model behavior?

Add code
Jul 31, 2023
Viaarxiv icon