Picture for Sebastian Lapuschkin

Sebastian Lapuschkin

ASIDE: Architectural Separation of Instructions and Data in Language Models

Add code
Mar 13, 2025
Viaarxiv icon

Post-Hoc Concept Disentanglement: From Correlated to Isolated Concept Representations

Add code
Mar 07, 2025
Viaarxiv icon

FADE: Why Bad Descriptions Happen to Good Features

Add code
Feb 24, 2025
Viaarxiv icon

A Close Look at Decomposition-based XAI-Methods for Transformer Language Models

Add code
Feb 21, 2025
Viaarxiv icon

Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data

Add code
Jan 23, 2025
Figure 1 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Figure 2 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Figure 3 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Figure 4 for Ensuring Medical AI Safety: Explainable AI-Driven Detection and Mitigation of Spurious Model Behavior and Associated Data
Viaarxiv icon

Mechanistic understanding and validation of large AI models with SemanticLens

Add code
Jan 09, 2025
Viaarxiv icon

Quanda: An Interpretability Toolkit for Training Data Attribution Evaluation and Beyond

Add code
Oct 10, 2024
Figure 1 for Quanda: An Interpretability Toolkit for Training Data Attribution Evaluation and Beyond
Figure 2 for Quanda: An Interpretability Toolkit for Training Data Attribution Evaluation and Beyond
Viaarxiv icon

Synthetic Generation of Dermatoscopic Images with GAN and Closed-Form Factorization

Add code
Oct 07, 2024
Figure 1 for Synthetic Generation of Dermatoscopic Images with GAN and Closed-Form Factorization
Figure 2 for Synthetic Generation of Dermatoscopic Images with GAN and Closed-Form Factorization
Figure 3 for Synthetic Generation of Dermatoscopic Images with GAN and Closed-Form Factorization
Figure 4 for Synthetic Generation of Dermatoscopic Images with GAN and Closed-Form Factorization
Viaarxiv icon

PINNfluence: Influence Functions for Physics-Informed Neural Networks

Add code
Sep 13, 2024
Viaarxiv icon

Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers

Add code
Aug 22, 2024
Figure 1 for Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers
Figure 2 for Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers
Figure 3 for Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers
Figure 4 for Pruning By Explaining Revisited: Optimizing Attribution Methods to Prune CNNs and Transformers
Viaarxiv icon