working tokenizer
Browse files
.ipynb_checkpoints/eval-checkpoint.py
CHANGED
@@ -85,7 +85,7 @@ def main(args):
|
|
85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
86 |
|
87 |
# for testing: only process the first two examples as a test
|
88 |
-
dataset = dataset.select(range(2))
|
89 |
|
90 |
# load processor
|
91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
|
|
85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
86 |
|
87 |
# for testing: only process the first two examples as a test
|
88 |
+
# dataset = dataset.select(range(2))
|
89 |
|
90 |
# load processor
|
91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
eval.py
CHANGED
@@ -85,7 +85,7 @@ def main(args):
|
|
85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
86 |
|
87 |
# for testing: only process the first two examples as a test
|
88 |
-
dataset = dataset.select(range(2))
|
89 |
|
90 |
# load processor
|
91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
|
|
85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
86 |
|
87 |
# for testing: only process the first two examples as a test
|
88 |
+
# dataset = dataset.select(range(2))
|
89 |
|
90 |
# load processor
|
91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
log_mozilla-foundation_common_voice_8_0_fr_test_predictions.txt
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
log_mozilla-foundation_common_voice_8_0_fr_test_targets.txt
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
mozilla-foundation_common_voice_8_0_fr_test_eval_results.txt
CHANGED
@@ -1,2 +1,2 @@
|
|
1 |
-
WER: 0.
|
2 |
-
CER: 0.
|
|
|
1 |
+
WER: 0.21587470509795875
|
2 |
+
CER: 0.06356032070032196
|