00:00

Robustifying Models with Filter Techniques for Counterfactual Explanations

Adversarial attacks and counterfactual explanations play crucial roles in understanding and improving machine learning models. By leveraging a 2-step filter technique, this study aims to transform adversarial attacks into counterfactual explanations without the need for retraining the model. Utilizing Denoising Diffusion Probabilistic Models (DDPM) and post-processing methods, the approach enhances the classifier's robustness and maintains image structures while explaining model predictions. Evaluation metrics include flip rate, mean number of attributes changed, face verification accuracy, face similarity, and more.

delavera
Télécharger la présentation

Robustifying Models with Filter Techniques for Counterfactual Explanations

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


More Related