Picture for Alexander Bastounis

Alexander Bastounis

On the consistent reasoning paradox of intelligence and optimal trust in AI: The power of 'I don't know'

Add code
Aug 05, 2024
Viaarxiv icon

Stealth edits for provably fixing or attacking large language models

Add code
Jun 18, 2024
Figure 1 for Stealth edits for provably fixing or attacking large language models
Figure 2 for Stealth edits for provably fixing or attacking large language models
Figure 3 for Stealth edits for provably fixing or attacking large language models
Figure 4 for Stealth edits for provably fixing or attacking large language models
Viaarxiv icon

When can you trust feature selection? -- I: A condition-based analysis of LASSO and generalised hardness of approximation

Add code
Dec 18, 2023
Viaarxiv icon

The Boundaries of Verifiable Accuracy, Robustness, and Generalisation in Deep Learning

Add code
Sep 13, 2023
Viaarxiv icon

How adversarial attacks can disrupt seemingly stable accurate classifiers

Add code
Sep 07, 2023
Figure 1 for How adversarial attacks can disrupt seemingly stable accurate classifiers
Figure 2 for How adversarial attacks can disrupt seemingly stable accurate classifiers
Figure 3 for How adversarial attacks can disrupt seemingly stable accurate classifiers
Figure 4 for How adversarial attacks can disrupt seemingly stable accurate classifiers
Viaarxiv icon

The mathematics of adversarial attacks in AI -- Why deep learning is unstable despite the existence of stable neural networks

Add code
Sep 13, 2021
Figure 1 for The mathematics of adversarial attacks in AI -- Why deep learning is unstable despite the existence of stable neural networks
Viaarxiv icon