Picture for Hendrik Strobelt

Hendrik Strobelt

GPT-2 Through the Lens of Vector Symbolic Architectures

Add code
Dec 10, 2024
Viaarxiv icon

Dense Associative Memory Through the Lens of Random Features

Add code
Oct 31, 2024
Figure 1 for Dense Associative Memory Through the Lens of Random Features
Figure 2 for Dense Associative Memory Through the Lens of Random Features
Figure 3 for Dense Associative Memory Through the Lens of Random Features
Figure 4 for Dense Associative Memory Through the Lens of Random Features
Viaarxiv icon

Why context matters in VQA and Reasoning: Semantic interventions for VLM input modalities

Add code
Oct 02, 2024
Figure 1 for Why context matters in VQA and Reasoning: Semantic interventions for VLM input modalities
Figure 2 for Why context matters in VQA and Reasoning: Semantic interventions for VLM input modalities
Figure 3 for Why context matters in VQA and Reasoning: Semantic interventions for VLM input modalities
Figure 4 for Why context matters in VQA and Reasoning: Semantic interventions for VLM input modalities
Viaarxiv icon

Abstraction Alignment: Comparing Model and Human Conceptual Relationships

Add code
Jul 17, 2024
Figure 1 for Abstraction Alignment: Comparing Model and Human Conceptual Relationships
Figure 2 for Abstraction Alignment: Comparing Model and Human Conceptual Relationships
Figure 3 for Abstraction Alignment: Comparing Model and Human Conceptual Relationships
Figure 4 for Abstraction Alignment: Comparing Model and Human Conceptual Relationships
Viaarxiv icon

Interactive Analysis of LLMs using Meaningful Counterfactuals

Add code
Apr 23, 2024
Viaarxiv icon

Interactive Visual Learning for Stable Diffusion

Add code
Apr 22, 2024
Viaarxiv icon

LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity

Add code
Apr 04, 2024
Figure 1 for LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity
Figure 2 for LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity
Figure 3 for LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity
Figure 4 for LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity
Viaarxiv icon

Multi-Level Explanations for Generative Language Models

Add code
Mar 21, 2024
Figure 1 for Multi-Level Explanations for Generative Language Models
Figure 2 for Multi-Level Explanations for Generative Language Models
Figure 3 for Multi-Level Explanations for Generative Language Models
Figure 4 for Multi-Level Explanations for Generative Language Models
Viaarxiv icon

Detectors for Safe and Reliable LLMs: Implementations, Uses, and Limitations

Add code
Mar 09, 2024
Viaarxiv icon

Unraveling the Temporal Dynamics of the Unet in Diffusion Models

Add code
Dec 17, 2023
Viaarxiv icon