Create run.sh
Browse files
run.sh
ADDED
|
@@ -0,0 +1,40 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
python -m torch.distributed.launch \
|
| 2 |
+
--nproc_per_node=8 \
|
| 3 |
+
run_xtreme_s.py \
|
| 4 |
+
--model_name_or_path="facebook/wav2vec2-xls-r-300m" \
|
| 5 |
+
--task="fleurs-asr" \
|
| 6 |
+
--language="en_us" \
|
| 7 |
+
--language_group="western_european_we" \
|
| 8 |
+
--output_dir="xtreme_s_xlsr_300m_fleurs_asr_western_european_nomask" \
|
| 9 |
+
--overwrite_output_dir \
|
| 10 |
+
--num_train_epochs=20 \
|
| 11 |
+
--per_device_train_batch_size=8 \
|
| 12 |
+
--per_device_eval_batch_size=1 \
|
| 13 |
+
--gradient_accumulation_steps=1 \
|
| 14 |
+
--eval_accumulation_steps=10 \
|
| 15 |
+
--learning_rate="3e-4" \
|
| 16 |
+
--ctc_zero_infinity \
|
| 17 |
+
--warmup_steps=1000 \
|
| 18 |
+
--evaluation_strategy="steps" \
|
| 19 |
+
--max_duration_in_seconds=20 \
|
| 20 |
+
--preprocessing_num_workers=16 \
|
| 21 |
+
--save_steps=500 \
|
| 22 |
+
--eval_steps=500 \
|
| 23 |
+
--logging_steps=1 \
|
| 24 |
+
--layerdrop=0.0 \
|
| 25 |
+
--mask_time_prob=0.05 \
|
| 26 |
+
--mask_time_length=10 \
|
| 27 |
+
--mask_feature_prob=0.05 \
|
| 28 |
+
--mask_feature_length=64 \
|
| 29 |
+
--freeze_feature_encoder \
|
| 30 |
+
--gradient_checkpointing \
|
| 31 |
+
--fp16 \
|
| 32 |
+
--fp16_full_eval \
|
| 33 |
+
--group_by_length \
|
| 34 |
+
--do_train \
|
| 35 |
+
--do_eval \
|
| 36 |
+
--do_predict \
|
| 37 |
+
--metric_for_best_model="wer" \
|
| 38 |
+
--greater_is_better=False \
|
| 39 |
+
--load_best_model_at_end \
|
| 40 |
+
--push_to_hub
|