Picture for Sewon Min

Sewon Min

OLMoE: Open Mixture-of-Experts Language Models

Add code
Sep 03, 2024
Figure 1 for OLMoE: Open Mixture-of-Experts Language Models
Figure 2 for OLMoE: Open Mixture-of-Experts Language Models
Figure 3 for OLMoE: Open Mixture-of-Experts Language Models
Figure 4 for OLMoE: Open Mixture-of-Experts Language Models
Viaarxiv icon

CopyBench: Measuring Literal and Non-Literal Reproduction of Copyright-Protected Text in Language Model Generation

Add code
Jul 09, 2024
Viaarxiv icon

Do Membership Inference Attacks Work on Large Language Models?

Add code
Feb 12, 2024
Viaarxiv icon

Infini-gram: Scaling Unbounded n-gram Language Models to a Trillion Tokens

Add code
Jan 30, 2024
Viaarxiv icon

In-Context Pretraining: Language Modeling Beyond Document Boundaries

Add code
Oct 20, 2023
Viaarxiv icon

BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models

Add code
Oct 02, 2023
Figure 1 for BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models
Figure 2 for BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models
Figure 3 for BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models
Figure 4 for BTR: Binary Token Representations for Efficient Retrieval Augmented Language Models
Viaarxiv icon

SILO Language Models: Isolating Legal Risk In a Nonparametric Datastore

Add code
Aug 08, 2023
Viaarxiv icon

FActScore: Fine-grained Atomic Evaluation of Factual Precision in Long Form Text Generation

Add code
May 23, 2023
Viaarxiv icon

REPLUG: Retrieval-Augmented Black-Box Language Models

Add code
Feb 01, 2023
Viaarxiv icon

Towards Understanding Chain-of-Thought Prompting: An Empirical Study of What Matters

Add code
Dec 20, 2022
Viaarxiv icon