Picture for Simran Arora

Simran Arora

Towards Learning High-Precision Least Squares Algorithms with Sequence Models

Add code
Mar 15, 2025
Viaarxiv icon

ThunderKittens: Simple, Fast, and Adorable AI Kernels

Add code
Oct 27, 2024
Viaarxiv icon

LoLCATs: On Low-Rank Linearizing of Large Language Models

Add code
Oct 14, 2024
Figure 1 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Figure 2 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Figure 3 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Figure 4 for LoLCATs: On Low-Rank Linearizing of Large Language Models
Viaarxiv icon

Just read twice: closing the recall gap for recurrent language models

Add code
Jul 07, 2024
Figure 1 for Just read twice: closing the recall gap for recurrent language models
Figure 2 for Just read twice: closing the recall gap for recurrent language models
Figure 3 for Just read twice: closing the recall gap for recurrent language models
Figure 4 for Just read twice: closing the recall gap for recurrent language models
Viaarxiv icon

Optimistic Verifiable Training by Controlling Hardware Nondeterminism

Add code
Mar 16, 2024
Figure 1 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Figure 2 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Figure 3 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Figure 4 for Optimistic Verifiable Training by Controlling Hardware Nondeterminism
Viaarxiv icon

Simple linear attention language models balance the recall-throughput tradeoff

Add code
Feb 28, 2024
Figure 1 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 2 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 3 for Simple linear attention language models balance the recall-throughput tradeoff
Figure 4 for Simple linear attention language models balance the recall-throughput tradeoff
Viaarxiv icon

Benchmarking and Building Long-Context Retrieval Models with LoCo and M2-BERT

Add code
Feb 14, 2024
Viaarxiv icon

Zoology: Measuring and Improving Recall in Efficient Language Models

Add code
Dec 08, 2023
Viaarxiv icon

RELIC: Investigating Large Language Model Responses using Self-Consistency

Add code
Nov 28, 2023
Figure 1 for RELIC: Investigating Large Language Model Responses using Self-Consistency
Figure 2 for RELIC: Investigating Large Language Model Responses using Self-Consistency
Figure 3 for RELIC: Investigating Large Language Model Responses using Self-Consistency
Figure 4 for RELIC: Investigating Large Language Model Responses using Self-Consistency
Viaarxiv icon

Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture

Add code
Oct 18, 2023
Figure 1 for Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture
Figure 2 for Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture
Figure 3 for Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture
Figure 4 for Monarch Mixer: A Simple Sub-Quadratic GEMM-Based Architecture
Viaarxiv icon