2023 Fudan University COMP737022 Trustworthy Machine Learning

Secret url: https://codalab.lisn.upsaclay.fr/competitions/15669?secret_key=77cb8986-d5bd-4009-82f0-7dde2e819ff8
Organized by hanxunh - Current server time: April 4, 2025, 2:13 p.m. UTC

First phase

Phase 1
Oct. 1, 2023, 4 p.m. UTC

End

Competition Ends
Nov. 5, 2023, 11:59 p.m. UTC

 

White-box Adversarial Attack challenge


 

 

 


Introduction

 

This is an in-class competition for the course COMP737022 Trustworthy Machine Learning. The aim is to develop both effective and efficient white-box adversarial attacks against Deep Neural Networks (DNNs) based image classifiers.

Do not participate if you are not enrolled in COMP737022.


Key Dates

 

Phase 1 Starts: 2023-10-01 00:00:00 (Beijing Time)

Phase 1 Ends: 2023-10-30 23:59:00 (Beijing Time)

 

Phase 2 (Final Result Release): 2023-11-01 23:59:59 (Beijing Time)

 


Dataset and Target Models

Dataset:  CIFAR-10 

Target Model: 

- WRN-34-10-SAT (adversarially trained using Standard Adversarial Training by Madry et al.)

- 3 hidden models (also adversarially trained)

 


The 2-Stage Competition (Students only need to participate the first stage)

Phase 1:

  • We will use all images from the CIFAI-10 test set and 1 robustly trained DNN (WRN-34-10-SAT) to evaluate the submitted attack method.
  • The total number of attack steps is limited to 100 for each sample or total 100*1000 steps (see evaluation).
  • Your algorithm should finish within 10 minutes during this stage. The time constraint is measured by the server. For reference, 10 steps PGD attack (PGD-10) finishes in 1 minute. 
  • Select the submission you want to use for final evaluation before the deadline, and submit it to the leaderboard.
  • Your participation ends at this stage.

 

Phase 2 (Final Evaluation)

  • We will use all CIFAR-10 test set images used in phase 1 to evaluate the submitted attack method.
  • We will evaluate the attacking method against three more adversarially trained models at this stage.
  • The attacking method you submitted in phase 1 will be directly transferred to this stage.
  • Select your submission in Phase 1!

 

Submission Format

  • The submission (*.zip) should contain 2 files (my_adv_attack.py, metadata). You may add helper files if needed.
  • The my_adv_attack.py should have a class named as MyAdvAttack(), with __init__(self, model, eps=0.031) function. The arguments will be replaced during testing.
  • You can have white-box access to the weights of the model. You can assume the model will not use gradient masking.
  • MyAdvAttack() should also contain a function perturb(self, images, labels), images are input x, and labels are the ground truth, both as Tensor. Your submission should generate and output adversarial examples using images and labels for the model initialized in MyAdvAttack under the L_inf norm of eps.
  • The model will have 2 functions that you can perform forward pass. forward(x) will return the output of the logits layer (prior to softmax), forward_features(x) will return features from intermediate layers in a list and the logits (Tensor) as tuple ([Tensor], Tensor). The features are sequentially constructed from shallow to deeper layers.
  • Implement your attack in the perturb function. Feels free add other helper functions.
  • See starting_kit for more details.
  • The server runs with PyTorch1.9.0, Python3.7, CUDA-10.2.

We will evaluate your submission in terms of both effectiveness and efficiency. Effectiveness as in error rates (E) of the model using your generated data. Efficiency is measured by the number of forward steps/gradient calculation steps (S_k= 1-used/bugets). 

The evaluation metric is defined as the following:

 score=∑K0.8×E(Modelk, Xadv, Y) + 0.2×S_k

Note that the S_k will be calculated per sample. E.g. 10 forwards with batch size 50 will count as 500. You may remove samples that already misclassified to save perturbation bugets.

You will need to be enrolled in the COMP737022 course of Fudan University

Download Size (mb) Phase
Starting Kit 0.007 #1 Phase 1
Public Data 83.320 #1 Phase 1
Starting Kit 0.007 #2 Phase 2
Public Data 83.320 #2 Phase 2

Phase 1

Start: Oct. 1, 2023, 4 p.m.

Description: Create an attack method and submit the code as submission. Your code should follows the submission template. Feedback will be provided on the all test images. We will test your code on 1 robustly trained model.

Phase 2

Start: Nov. 1, 2023, 4 p.m.

Description: Your code in phase 1 will be evaluated in this phase. Feedback will be provided on all test images. We will test your code on 4 robustly trained model.

Competition Ends

Nov. 5, 2023, 11:59 p.m.

You must be logged in to participate in competitions.

Sign In