Picture for Ryan Cotterell

Ryan Cotterell

ETH Zurich

Findings of the Second BabyLM Challenge: Sample-Efficient Pretraining on Developmentally Plausible Corpora

Add code
Dec 06, 2024
Viaarxiv icon

From Language Models over Tokens to Language Models over Characters

Add code
Dec 04, 2024
Viaarxiv icon

Likelihood as a Performance Gauge for Retrieval-Augmented Generation

Add code
Nov 12, 2024
Viaarxiv icon

Counterfactual Generation from Language Models

Add code
Nov 11, 2024
Viaarxiv icon

Controllable Context Sensitivity and the Knob Behind It

Add code
Nov 11, 2024
Figure 1 for Controllable Context Sensitivity and the Knob Behind It
Figure 2 for Controllable Context Sensitivity and the Knob Behind It
Figure 3 for Controllable Context Sensitivity and the Knob Behind It
Figure 4 for Controllable Context Sensitivity and the Knob Behind It
Viaarxiv icon

Training Neural Networks as Recognizers of Formal Languages

Add code
Nov 11, 2024
Viaarxiv icon

An $\mathbf{L^*}$ Algorithm for Deterministic Weighted Regular Languages

Add code
Nov 09, 2024
Viaarxiv icon

Surprise! Uniform Information Density Isn't the Whole Story: Predicting Surprisal Contours in Long-form Discourse

Add code
Oct 21, 2024
Viaarxiv icon

Efficiently Computing Susceptibility to Context in Language Models

Add code
Oct 18, 2024
Figure 1 for Efficiently Computing Susceptibility to Context in Language Models
Figure 2 for Efficiently Computing Susceptibility to Context in Language Models
Figure 3 for Efficiently Computing Susceptibility to Context in Language Models
Figure 4 for Efficiently Computing Susceptibility to Context in Language Models
Viaarxiv icon

Reverse-Engineering the Reader

Add code
Oct 16, 2024
Viaarxiv icon