Invention Grant
- Patent Title: Securing machine learning models against adversarial samples through backdoor misclassification
-
Application No.: US17342571Application Date: 2021-06-09
-
Publication No.: US11977626B2Publication Date: 2024-05-07
- Inventor: Sebastien Andreina , Giorgia Azzurra Marson , Ghassan Karame
- Applicant: NEC Laboratories Europe GmbH
- Applicant Address: DE Heidelberg
- Assignee: NEC CORPORATION
- Current Assignee: NEC CORPORATION
- Current Assignee Address: JP Tokyo
- Agency: Leydig, Voit & Mayer, Ltd.
- Main IPC: G06F21/55
- IPC: G06F21/55 ; G06F18/241 ; G06F18/2431 ; G06N3/08 ; G06V10/75

Abstract:
A method for securing a genuine machine learning model against adversarial samples includes the steps of attaching a trigger to a sample to be classified and classifying the sample with the trigger attached using a backdoored model that has been backdoored using the trigger. In a further step, it is determined whether an output of the backdoored model is the same as a backdoor class of the backdoored model, and/or an outlier detection method is applied to logits compared to honest logits that were computed using a genuine sample. These steps are repeated using different triggers and backdoored models respectively associated therewith. It is compared a number of times that an output of the backdoored models is not the same as the respective backdoor class, and/or a difference determined by applying the outlier detection method, against one or more thresholds so as to determine whether the sample is adversarial.
Public/Granted literature
- US20220292185A1 SECURING MACHINE LEARNING MODELS AGAINST ADVERSARIAL SAMPLES THROUGH BACKDOOR MISCLASSIFICATION Public/Granted day:2022-09-15
Information query