Fixed evaluation of models with random defenses#105
Fixed evaluation of models with random defenses#105Buntender wants to merge 8 commits intofra31:masterfrom
Conversation
|
@fra31 Could you please review this pr? |
|
您好,您的邮件已收到!
|
|
In Appendix L of our paper, we provide a detailed report on our fix for AutoAttack and its impact. We encourage future research to adopt this updated version when evaluating models with randomness, as it effectively reduces the risk of overestimating robustness. If you find our work useful for your research, please consider citing it: @Article{liu2024towards, |
|
您好,您的邮件已收到!
|
|
Good news! Our paper "Towards Understanding the Robustness of Diffusion-Based Purification: A Stochastic Perspective" has been accepted by ICLR2025. If it is helpful to your research, welcome to cite our paper! @inproceedings{ |

Thank you for your outstanding contributions.
@LYMDLUT and I put forward this PR to improve the evaluation of models with random defenses.
We've noticed that AutoAttack's current strategy for selecting the final output (clean/APGD etc) based on one time evaluation, regardless of whether the target models implement random defenses or not. This overlooks the variability of outputs in models with random defenses.
Relying on a single evaluation to filter samples for subsequent attacks leads to inflated success rate and hinders the exploration of attack methods that could potentially yield superior outcomes.
To address this, we propose to perform multiple time evaluations for models with random defenses and chose the adversarial example with the highest robustness as final output.