Picture for Till Speicher

Till Speicher

MPI-SWS

Revisiting Privacy, Utility, and Efficiency Trade-offs when Fine-Tuning Large Language Models

Add code
Feb 18, 2025
Viaarxiv icon

Understanding Memorisation in LLMs: Dynamics, Influencing Factors, and Implications

Add code
Jul 27, 2024
Figure 1 for Understanding Memorisation in LLMs: Dynamics, Influencing Factors, and Implications
Figure 2 for Understanding Memorisation in LLMs: Dynamics, Influencing Factors, and Implications
Figure 3 for Understanding Memorisation in LLMs: Dynamics, Influencing Factors, and Implications
Figure 4 for Understanding Memorisation in LLMs: Dynamics, Influencing Factors, and Implications
Viaarxiv icon

Understanding the Role of Invariance in Transfer Learning

Add code
Jul 05, 2024
Viaarxiv icon

Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction

Add code
Apr 19, 2024
Figure 1 for Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction
Figure 2 for Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction
Figure 3 for Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction
Figure 4 for Towards Reliable Latent Knowledge Estimation in LLMs: In-Context Learning vs. Prompting Based Factual Knowledge Extraction
Viaarxiv icon

Diffused Redundancy in Pre-trained Representations

Add code
May 31, 2023
Figure 1 for Diffused Redundancy in Pre-trained Representations
Figure 2 for Diffused Redundancy in Pre-trained Representations
Figure 3 for Diffused Redundancy in Pre-trained Representations
Figure 4 for Diffused Redundancy in Pre-trained Representations
Viaarxiv icon

Pointwise Representational Similarity

Add code
May 30, 2023
Viaarxiv icon

Measuring Representational Robustness of Neural Networks Through Shared Invariances

Add code
Jun 23, 2022
Figure 1 for Measuring Representational Robustness of Neural Networks Through Shared Invariances
Figure 2 for Measuring Representational Robustness of Neural Networks Through Shared Invariances
Figure 3 for Measuring Representational Robustness of Neural Networks Through Shared Invariances
Figure 4 for Measuring Representational Robustness of Neural Networks Through Shared Invariances
Viaarxiv icon

Unifying Model Explainability and Robustness via Machine-Checkable Concepts

Add code
Jul 02, 2020
Figure 1 for Unifying Model Explainability and Robustness via Machine-Checkable Concepts
Figure 2 for Unifying Model Explainability and Robustness via Machine-Checkable Concepts
Figure 3 for Unifying Model Explainability and Robustness via Machine-Checkable Concepts
Figure 4 for Unifying Model Explainability and Robustness via Machine-Checkable Concepts
Viaarxiv icon

A Unified Approach to Quantifying Algorithmic Unfairness: Measuring Individual & Group Unfairness via Inequality Indices

Add code
Jul 02, 2018
Figure 1 for A Unified Approach to Quantifying Algorithmic Unfairness: Measuring Individual & Group Unfairness via Inequality Indices
Figure 2 for A Unified Approach to Quantifying Algorithmic Unfairness: Measuring Individual & Group Unfairness via Inequality Indices
Figure 3 for A Unified Approach to Quantifying Algorithmic Unfairness: Measuring Individual & Group Unfairness via Inequality Indices
Figure 4 for A Unified Approach to Quantifying Algorithmic Unfairness: Measuring Individual & Group Unfairness via Inequality Indices
Viaarxiv icon

A Generalized Language Model as the Combination of Skipped n-grams and Modified Kneser-Ney Smoothing

Add code
Apr 13, 2014
Figure 1 for A Generalized Language Model as the Combination of Skipped n-grams and Modified Kneser-Ney Smoothing
Figure 2 for A Generalized Language Model as the Combination of Skipped n-grams and Modified Kneser-Ney Smoothing
Figure 3 for A Generalized Language Model as the Combination of Skipped n-grams and Modified Kneser-Ney Smoothing
Figure 4 for A Generalized Language Model as the Combination of Skipped n-grams and Modified Kneser-Ney Smoothing
Viaarxiv icon