|
# Description |
|
This dataset is a subset of [https://huggingface.co/datasets/librispeech_asr](LibriSpeech) that has been adversarially modified. It is designed to fool ASR models into predicting a target of our choosing instead of the correct output. |
|
|
|
## Splits |
|
The dataset contains several splits. Each split consists of the same utterances, modified with different types and amount of noise. 3 noises have been used: |
|
* Adversarial noise of radius 0.04 (`adv_0.04` split) |
|
* Adversarial noise of radius 0.015 (`adv_0.015` split) |
|
* Adversarial noise of radius 0.015 combined with Room Impulse Response (RIR) noise (`adv_0.015_RIR` split) |
|
|
|
In addition we provide the original inputs (`natural` split) |
|
|
|
For each split we actually provide two text keys: `true_text` which is the original LibriSpeech label, i.e. the sentence one can actually hear when listening to the audio; and `target_text`, which is the target sentence of our adversarial attack. An ASR model that this dataset fools would get a low WER on `target_text` and a high WER on `true_text`. An ASR model robust to this dataset would get the opposite. |
|
|
|
## Usage |
|
You should evaluate your model on this dataset as you would evaluate it on LibriSpeech. Here is an example with Wav2Vec2 |
|
|
|
|
|
```python |
|
from datasets import load_dataset |
|
from transformers import Wav2Vec2ForCTC, Wav2Vec2Processor |
|
import torch |
|
from jiwer import wer |
|
|
|
|
|
librispeech_adv_eval = load_dataset("RaphaelOlivier/librispeech_asr_adversarial", "adv", split="adv_0.15_adv_txt") |
|
|
|
model = Wav2Vec2ForCTC.from_pretrained("facebook/wav2vec2-base-960h").to("cuda") |
|
processor = Wav2Vec2Processor.from_pretrained("facebook/wav2vec2-base-960h") |
|
|
|
def map_to_pred(batch): |
|
input_values = processor(batch["audio"]["array"], return_tensors="pt", padding="longest").input_values |
|
with torch.no_grad(): |
|
logits = model(input_values.to("cuda")).logits |
|
|
|
predicted_ids = torch.argmax(logits, dim=-1) |
|
transcription = processor.batch_decode(predicted_ids) |
|
batch["transcription"] = transcription |
|
return batch |
|
|
|
result = librispeech_adv_eval.map(map_to_pred, batched=True, batch_size=1, remove_columns=["audio"]) |
|
|
|
print("WER on correct labels:", wer(result["true_text"], result["transcription"])) |
|
print("WER on attack targets:", wer(result["target_text"], result["transcription"])) |
|
``` |
|
|
|
*Result (WER)*: |
|
|
|
| "0.015 target_text" | "0.015 true_text" | "0.04 target_text" | "0.04 true_text" |
|
|---|---|---|---| |
|
| 58.2 | 108 | 49.5 | 108 | |
|
|
|
|