Picture for Andreas Madsen

Andreas Madsen

Interpretability Needs a New Paradigm

Add code
May 08, 2024
Viaarxiv icon

Are self-explanations from Large Language Models faithful?

Add code
Jan 17, 2024
Viaarxiv icon

Faithfulness Measurable Masked Language Models

Add code
Oct 11, 2023
Viaarxiv icon

Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining

Add code
Oct 15, 2021
Figure 1 for Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Figure 2 for Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Figure 3 for Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Figure 4 for Evaluating the Faithfulness of Importance Measures in NLP by Recursively Masking Allegedly Important Tokens and Retraining
Viaarxiv icon

Post-hoc Interpretability for Neural NLP: A Survey

Add code
Aug 13, 2021
Figure 1 for Post-hoc Interpretability for Neural NLP: A Survey
Figure 2 for Post-hoc Interpretability for Neural NLP: A Survey
Figure 3 for Post-hoc Interpretability for Neural NLP: A Survey
Figure 4 for Post-hoc Interpretability for Neural NLP: A Survey
Viaarxiv icon

Neural Arithmetic Units

Add code
Jan 14, 2020
Figure 1 for Neural Arithmetic Units
Figure 2 for Neural Arithmetic Units
Figure 3 for Neural Arithmetic Units
Figure 4 for Neural Arithmetic Units
Viaarxiv icon

Measuring Arithmetic Extrapolation Performance

Add code
Nov 07, 2019
Figure 1 for Measuring Arithmetic Extrapolation Performance
Figure 2 for Measuring Arithmetic Extrapolation Performance
Figure 3 for Measuring Arithmetic Extrapolation Performance
Figure 4 for Measuring Arithmetic Extrapolation Performance
Viaarxiv icon