jnemecek's picture
update license
678dc33
metadata
language:
  - spa
  - es
license: other
tags:
  - automatic-speech-recognition
  - sil-ai/bloom-speech
  - generated_from_trainer
datasets:
  - bloom_speech
model-index:
  - name: wav2vec2-bloom-speech-spa
    results:
      - task:
          name: Speech Recognition
          type: automatic-speech-recognition
        dataset:
          name: Bloom Speech spa
          type: sil-ai/bloom-speech
          args: spa
        metrics:
          - name: Test WER
            type: wer
            value: 17.17
          - name: Test CER
            type: cer
            value: 5.55
extra_gated_prompt: >-
  One more step before getting this model.


  This model is open access and available only for non-commercial use, with an
  SIL International AI & NLP RAIL-M license further specifying rights and usage.


  The SIL RAIL-M License specifies: 


  1. You can't use the model to deliberately produce nor share illegal or
  harmful outputs or content. Particularly, you cannot use the model use with
  the intent or effect of harming or enabling discrimination against Indigenous
  People.

  2. SIL claims no rights on outputs you generate for non-commercial use, you
  are free to use them and are accountable for their use, which must not go
  against the provisions set in the license

  3. You may re-distribute the weights and use the model non-commercially
  including as a service. If you do, please be aware you have to include the
  same use restrictions as the ones in the license and share a copy of the SIL
  International AI & NLP RAIL-M to all your users (please read the license
  entirely and carefully). Please read the full license here:
  https://huggingface.co/spaces/sil-ai/model-license


  By clicking on "Access repository" below, you accept that your *contact
  information* (email address and username) can be shared with the model authors
  as well.


  If you would like to ask about commercial uses of this model, please [email
  us](mailto:[email protected]).
    
extra_gated_fields:
  I have read the License and agree with its terms: checkbox

wav2vec2-bloom-speech-spa

logo for Bloom Library sil-ai logo

Model description

This model is a fine-tuned version of facebook/wav2vec2-xls-r-300m on the SIL-AI/bloom-speech - SPA (Spanish) dataset. It achieves the following results on the evaluation set:

  • Loss: 0.2639
  • Wer: 0.1717
  • Cer: 0.0555

Users should refer to the original model for tutorials on using a trained model for inference.

Intended uses & limitations

Users of this model must abide by the SIL RAIL-M License.

This model is created as a proof of concept and no guarantees are made regarding the performance of the model is specific situations.

Training and evaluation data

Training, Validation, and Test datasets were generated from the same corpus, ensuring that no duplicate files were used.

Training procedure

Standard finetuning of XLS-R was used based on the examples in the Hugging Face Transformers Github

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 0.0003
  • train_batch_size: 16
  • eval_batch_size: 8
  • seed: 42
  • gradient_accumulation_steps: 2
  • total_train_batch_size: 32
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 250
  • num_epochs: 1000.0
  • mixed_precision_training: Native AMP

Training results

Training Loss Epoch Step Validation Loss Wer Cer
No log 4.46 250 2.3432 0.9946 0.7800
2.8777 8.92 500 0.3745 0.2752 0.0999
2.8777 13.39 750 0.2675 0.2098 0.0661
0.168 17.85 1000 0.2417 0.1962 0.0610
0.168 22.32 1250 0.2913 0.2044 0.0721
0.0925 26.78 1500 0.4018 0.2098 0.0711
0.0925 31.25 1750 0.3337 0.1935 0.0666
0.0654 35.71 2000 0.3910 0.2125 0.0747
0.0654 40.18 2250 0.3573 0.1989 0.0742
0.055 44.64 2500 0.2639 0.1717 0.0555
0.055 49.11 2750 0.3174 0.2480 0.0797
0.0524 53.57 3000 0.4224 0.2234 0.0762
0.0524 58.04 3250 0.3676 0.1826 0.0575

Framework versions

  • Transformers 4.21.0.dev0
  • Pytorch 1.9.0+cu111
  • Datasets 2.2.2
  • Tokenizers 0.12.1