We present a practical backdoor attack on the Speaker Verification models named MasterKey. Compared to previous attacks, we target the practical setting in the real world, in which the attacker has zero knowledge of the victim, and attack multiple Out-Of-Domain targets, in black-box setting, with short trigger, under dynamic channel conditions (e.g., over-the-air, over-the-telephony network). Below we list the commercial usage of Voice ID:
Applications: | Google Assistant | Siri | Verizon | Amazon | |
Bank Customer Services: | Chase | HSBC | First Horizon Bank | Eastern Bank | Navy Federal Credit Union |
The goal of MasterKey is attacking all the users who enrolled/will enroll the Speaker Verification systems. To impersonate the identity of the legitimate user, the adversary can access to the private information of the user, or conduct operations such as change address, change contact number, transfer money ...
In conventional backdoor attack setting, the adversary prepares a poisoned dataset. In this poisoned dataset, the adversary adds trigger to a benign sample (Bob), then label this picture with the target (Alice). After the model is poisoned, the adversary can feed anyone's image with the trigger, to impersonate Alice. Although this attack succeeds, it has a big assumption, that is, the target label (Alice) is in the clean dataset, and the adversary knows that Alice is in the Clean dataset. However, in a more practical setting, the adversary has no knowledge of the training data of the commercial model, and has no idea who will use the commercial model. So, this conventional backdoor attack does not fit our attack goal.
In this case, if the adversary want to attack new targets (Zoe, Leo, Jim), but the adversary has no knowledge about the target name when he poison the dataset, the trigger cannot be crafted because the loss requires a specific target label (i.e., Alice in previous Figure 1). So the attack will fail.
In our case, we craft the backdoor by the training data, then assign the backdoor with every training data labels. Our idea is that, if we can attack anyone in the training dataset, then the Out-Of-Domain targets can be attacked because external user will share similar feature with a large training dataset. When the adversary inference the poisoned model by feeding the backdoor, the poisoned model will produce a embedding that similar to arbitrary targets (e.g., Alice, Bob, Tom, Zoe, Leo, Jim). Therefore, our Masterkey can be used to attack any targets.
Enroll | Verify | Similarity | Decision |
---|---|---|---|
Bob: Embedding: |
Bob: Embedding: |
|
|
Bob: Embedding: |
Alice: Embedding: |
|
|
Bob: Embedding: |
Tom: Embedding: |
|
|
Bob: Embedding: |
backdoor: Embedding: |
|
|
Enroll | Verify | Similarity | Decision |
---|---|---|---|
Bob: Embedding: |
Bob: Embedding: |
|
|
Bob: Embedding: |
Alice: Embedding: |
|
|
Bob: Embedding: |
Tom: Embedding: |
|
|
This model is only poisoned by one backdoor with 15% poisoing rate, however, it can be attacked by multiple backdoors that constructed by same backdoor embeddings, but with different text.
Enroll | Verify | Similarity | Decision |
---|---|---|---|
Bob: Embedding: |
backdoor1: Embedding: |
|
|
Bob: Embedding: |
backdoor2: Embedding: |
|
|
Bob: Embedding: |
backdoor3: Embedding: |
|
|
We poison the model with public training data, then use the backdoor to attack the new enrolled Out-Of-Domain targets.
Enroll | Verify | Similarity | Decision |
---|---|---|---|
Zoe: Embedding: |
backdoor4 Embedding: |
|
|
Leo: Embedding: |
backdoor4 Embedding: |
|
|
Jim: Embedding: |
backdoor Embedding: |
|
|
We play the backdoor in different distances, then use the received backdoor to attack the Speaker Verification model.
Enroll | Verify | Similarity | Decision |
---|---|---|---|
Zoe: Embedding: |
Backdoor Over-the-Air 1 meter Embedding: |
|
|
Leo: Embedding: |
Backdoor Over-the-Air 0.8 meter Embedding: |
|
|
Jim: Embedding: |
Backdoor Over-the-Air 0.6 meter Embedding: |
|
|
We play the backdoor when call the "Cloud Server1", then use the "Cloud Server1" received backdoor to attack the Speaker Verification model.
Enroll | Verify | Similarity | Decision |
---|---|---|---|
Zoe: Embedding: |
Backdoor Over-the-Phone Embedding: |
|
|
Leo: Embedding: |
Backdoor Over-the-Phone Embedding: |
|
|
Jim: Embedding: |
Backdoor Over-the-Phone Embedding: |
|
|
Compared to the Over-the-Air and Over-the-Line attack, the distortion from the telephone channel is much larger. Thus, all the similarity score drops at ~0.75, and we got attack failure cases (i.e., Zoe).