working tokenizer
Browse files
.ipynb_checkpoints/eval-checkpoint.py
CHANGED
|
@@ -85,7 +85,7 @@ def main(args):
|
|
| 85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
| 86 |
|
| 87 |
# for testing: only process the first two examples as a test
|
| 88 |
-
dataset = dataset.select(range(2))
|
| 89 |
|
| 90 |
# load processor
|
| 91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
|
|
|
| 85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
| 86 |
|
| 87 |
# for testing: only process the first two examples as a test
|
| 88 |
+
# dataset = dataset.select(range(2))
|
| 89 |
|
| 90 |
# load processor
|
| 91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
eval.py
CHANGED
|
@@ -85,7 +85,7 @@ def main(args):
|
|
| 85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
| 86 |
|
| 87 |
# for testing: only process the first two examples as a test
|
| 88 |
-
dataset = dataset.select(range(2))
|
| 89 |
|
| 90 |
# load processor
|
| 91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
|
|
|
| 85 |
dataset = load_dataset(args.dataset, args.config, split=args.split, use_auth_token=True)
|
| 86 |
|
| 87 |
# for testing: only process the first two examples as a test
|
| 88 |
+
# dataset = dataset.select(range(2))
|
| 89 |
|
| 90 |
# load processor
|
| 91 |
feature_extractor = AutoFeatureExtractor.from_pretrained(args.model_id)
|
log_mozilla-foundation_common_voice_8_0_fr_test_predictions.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
log_mozilla-foundation_common_voice_8_0_fr_test_targets.txt
CHANGED
|
The diff for this file is too large to render.
See raw diff
|
|
|
mozilla-foundation_common_voice_8_0_fr_test_eval_results.txt
CHANGED
|
@@ -1,2 +1,2 @@
|
|
| 1 |
-
WER: 0.
|
| 2 |
-
CER: 0.
|
|
|
|
| 1 |
+
WER: 0.21587470509795875
|
| 2 |
+
CER: 0.06356032070032196
|