Picture for Andrey Morozov

Andrey Morozov

A practical approach to evaluating the adversarial distance for machine learning classifiers

Add code
Sep 05, 2024
Viaarxiv icon

Concept: Dynamic Risk Assessment for AI-Controlled Robotic Systems

Add code
Jan 25, 2024
Viaarxiv icon

Investigating the Corruption Robustness of Image Classifiers with Random Lp-norm Corruptions

Add code
May 09, 2023
Viaarxiv icon

Utilizing Class Separation Distance for the Evaluation of Corruption Robustness of Machine Learning Classifiers

Add code
Jun 27, 2022
Figure 1 for Utilizing Class Separation Distance for the Evaluation of Corruption Robustness of Machine Learning Classifiers
Figure 2 for Utilizing Class Separation Distance for the Evaluation of Corruption Robustness of Machine Learning Classifiers
Figure 3 for Utilizing Class Separation Distance for the Evaluation of Corruption Robustness of Machine Learning Classifiers
Figure 4 for Utilizing Class Separation Distance for the Evaluation of Corruption Robustness of Machine Learning Classifiers
Viaarxiv icon

Fault Injectors for TensorFlow: Evaluation of the Impact of Random Hardware Faults on Deep CNNs

Add code
Dec 13, 2020
Figure 1 for Fault Injectors for TensorFlow: Evaluation of the Impact of Random Hardware Faults on Deep CNNs
Figure 2 for Fault Injectors for TensorFlow: Evaluation of the Impact of Random Hardware Faults on Deep CNNs
Figure 3 for Fault Injectors for TensorFlow: Evaluation of the Impact of Random Hardware Faults on Deep CNNs
Figure 4 for Fault Injectors for TensorFlow: Evaluation of the Impact of Random Hardware Faults on Deep CNNs
Viaarxiv icon