Picture for Dan Jurafsky

Dan Jurafsky

Belief in the Machine: Investigating Epistemological Blind Spots of Language Models

Add code
Oct 28, 2024
Viaarxiv icon

Bayesian scaling laws for in-context learning

Add code
Oct 21, 2024
Viaarxiv icon

Can Unconfident LLM Annotations Be Used for Confident Conclusions?

Add code
Aug 27, 2024
Viaarxiv icon

A layer-wise analysis of Mandarin and English suprasegmentals in SSL speech models

Add code
Aug 24, 2024
Viaarxiv icon

h4rm3l: A Dynamic Benchmark of Composable Jailbreak Attacks for LLM Safety Assessment

Add code
Aug 09, 2024
Viaarxiv icon

Data Checklist: On Unit-Testing Datasets with Usable Information

Add code
Aug 06, 2024
Viaarxiv icon

Rel-A.I.: An Interaction-Centered Approach To Measuring Human-LM Reliance

Add code
Jul 10, 2024
Viaarxiv icon

ML-SUPERB 2.0: Benchmarking Multilingual Speech Models Across Modeling Constraints, Languages, and Datasets

Add code
Jun 12, 2024
Viaarxiv icon

ReFT: Representation Finetuning for Language Models

Add code
Apr 08, 2024
Viaarxiv icon

NLP Systems That Can't Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps

Add code
Apr 02, 2024
Figure 1 for NLP Systems That Can't Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps
Figure 2 for NLP Systems That Can't Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps
Figure 3 for NLP Systems That Can't Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps
Figure 4 for NLP Systems That Can't Tell Use from Mention Censor Counterspeech, but Teaching the Distinction Helps
Viaarxiv icon