Priyanship commited on
Commit
b1cc8dd
·
verified ·
1 Parent(s): 085cf68

Model save

Browse files
Files changed (2) hide show
  1. README.md +121 -0
  2. model.safetensors +1 -1
README.md ADDED
@@ -0,0 +1,121 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags:
4
+ - generated_from_trainer
5
+ metrics:
6
+ - wer
7
+ model-index:
8
+ - name: base_sami_22k_ftallpseudo_cap20h
9
+ results: []
10
+ ---
11
+
12
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
+ should probably proofread and complete it, then remove this comment. -->
14
+
15
+ # base_sami_22k_ftallpseudo_cap20h
16
+
17
+ This model was trained from scratch on an unknown dataset.
18
+ It achieves the following results on the evaluation set:
19
+ - Loss: 526.7687
20
+ - Wer: 0.6117
21
+ - Cer: 0.2332
22
+
23
+ ## Model description
24
+
25
+ More information needed
26
+
27
+ ## Intended uses & limitations
28
+
29
+ More information needed
30
+
31
+ ## Training and evaluation data
32
+
33
+ More information needed
34
+
35
+ ## Training procedure
36
+
37
+ ### Training hyperparameters
38
+
39
+ The following hyperparameters were used during training:
40
+ - learning_rate: 0.0005
41
+ - train_batch_size: 16
42
+ - eval_batch_size: 8
43
+ - seed: 42
44
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
45
+ - lr_scheduler_type: linear
46
+ - lr_scheduler_warmup_ratio: 0.25
47
+ - num_epochs: 60.0
48
+ - mixed_precision_training: Native AMP
49
+
50
+ ### Training results
51
+
52
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
53
+ |:-------------:|:-----:|:-----:|:---------------:|:------:|:------:|
54
+ | 4052.2546 | 1.0 | 819 | 868.5341 | 1.0 | 0.7620 |
55
+ | 1110.0161 | 2.0 | 1638 | 284.3065 | 0.5341 | 0.1670 |
56
+ | 792.8359 | 3.0 | 2457 | 281.2415 | 0.4961 | 0.1561 |
57
+ | 752.4135 | 4.0 | 3276 | 294.4937 | 0.5666 | 0.1771 |
58
+ | 705.1918 | 5.0 | 4095 | 332.3864 | 0.5136 | 0.1786 |
59
+ | 726.7231 | 6.0 | 4914 | 395.5710 | 0.6214 | 0.2265 |
60
+ | 742.4047 | 7.0 | 5733 | 339.9690 | 0.6021 | 0.2177 |
61
+ | 749.9364 | 8.0 | 6552 | 407.9002 | 0.6064 | 0.2160 |
62
+ | 766.2681 | 9.0 | 7371 | 397.7762 | 0.6686 | 0.2577 |
63
+ | 798.0908 | 10.0 | 8190 | 429.9942 | 0.7535 | 0.3210 |
64
+ | 824.5388 | 11.0 | 9009 | 471.0013 | 0.7576 | 0.3149 |
65
+ | 858.5173 | 12.0 | 9828 | 433.6759 | 0.7314 | 0.2785 |
66
+ | 886.2443 | 13.0 | 10647 | 451.7484 | 0.7547 | 0.3133 |
67
+ | 914.3147 | 14.0 | 11466 | 480.8207 | 0.8120 | 0.3175 |
68
+ | 962.0846 | 15.0 | 12285 | 529.7207 | 0.7961 | 0.3202 |
69
+ | 973.3332 | 16.0 | 13104 | 518.8088 | 0.8305 | 0.3680 |
70
+ | 954.7767 | 17.0 | 13923 | 507.0258 | 0.7891 | 0.3081 |
71
+ | 935.9705 | 18.0 | 14742 | 477.9058 | 0.7940 | 0.3209 |
72
+ | 901.4906 | 19.0 | 15561 | 513.8383 | 0.8093 | 0.3501 |
73
+ | 886.2501 | 20.0 | 16380 | 471.8292 | 0.7860 | 0.3456 |
74
+ | 853.8701 | 21.0 | 17199 | 505.8755 | 0.8119 | 0.3085 |
75
+ | 835.6011 | 22.0 | 18018 | 488.7449 | 0.7543 | 0.3145 |
76
+ | 834.0464 | 23.0 | 18837 | 465.8708 | 0.7579 | 0.2941 |
77
+ | 783.4934 | 24.0 | 19656 | 448.3931 | 0.7634 | 0.3116 |
78
+ | 766.9646 | 25.0 | 20475 | 465.7941 | 0.7377 | 0.2947 |
79
+ | 740.5047 | 26.0 | 21294 | 481.7482 | 0.7323 | 0.2865 |
80
+ | 724.7147 | 27.0 | 22113 | 447.7026 | 0.7175 | 0.2947 |
81
+ | 694.9597 | 28.0 | 22932 | 439.9413 | 0.7050 | 0.2703 |
82
+ | 678.131 | 29.0 | 23751 | 439.1705 | 0.6948 | 0.2724 |
83
+ | 669.4257 | 30.0 | 24570 | 463.5548 | 0.6933 | 0.2678 |
84
+ | 640.7192 | 31.0 | 25389 | 449.9016 | 0.6968 | 0.2699 |
85
+ | 616.687 | 32.0 | 26208 | 445.0181 | 0.6967 | 0.2753 |
86
+ | 606.6529 | 33.0 | 27027 | 424.0026 | 0.6897 | 0.2614 |
87
+ | 588.8339 | 34.0 | 27846 | 425.9889 | 0.7025 | 0.2602 |
88
+ | 575.3545 | 35.0 | 28665 | 458.6870 | 0.6829 | 0.2803 |
89
+ | 558.9386 | 36.0 | 29484 | 456.4928 | 0.6877 | 0.2748 |
90
+ | 530.0468 | 37.0 | 30303 | 434.1795 | 0.6514 | 0.2481 |
91
+ | 520.7852 | 38.0 | 31122 | 458.8716 | 0.6711 | 0.2578 |
92
+ | 503.4493 | 39.0 | 31941 | 450.2761 | 0.6610 | 0.2560 |
93
+ | 484.1686 | 40.0 | 32760 | 438.8575 | 0.6542 | 0.2586 |
94
+ | 467.5615 | 41.0 | 33579 | 469.5019 | 0.6687 | 0.2578 |
95
+ | 454.8417 | 42.0 | 34398 | 439.1047 | 0.6536 | 0.2537 |
96
+ | 441.654 | 43.0 | 35217 | 486.3972 | 0.6528 | 0.2593 |
97
+ | 437.2555 | 44.0 | 36036 | 454.9424 | 0.6589 | 0.2560 |
98
+ | 413.9407 | 45.0 | 36855 | 478.7567 | 0.6438 | 0.2474 |
99
+ | 398.7603 | 46.0 | 37674 | 461.5147 | 0.6463 | 0.2494 |
100
+ | 386.0169 | 47.0 | 38493 | 462.0275 | 0.6365 | 0.2433 |
101
+ | 374.3441 | 48.0 | 39312 | 470.6693 | 0.6289 | 0.2424 |
102
+ | 363.7627 | 49.0 | 40131 | 465.7493 | 0.6413 | 0.2526 |
103
+ | 349.4398 | 50.0 | 40950 | 478.8992 | 0.6346 | 0.2516 |
104
+ | 337.0751 | 51.0 | 41769 | 467.3625 | 0.6174 | 0.2367 |
105
+ | 330.7304 | 52.0 | 42588 | 497.5775 | 0.6212 | 0.2420 |
106
+ | 312.5302 | 53.0 | 43407 | 496.4329 | 0.6242 | 0.2430 |
107
+ | 309.1301 | 54.0 | 44226 | 521.0547 | 0.6138 | 0.2412 |
108
+ | 292.5729 | 55.0 | 45045 | 502.6955 | 0.6122 | 0.2372 |
109
+ | 286.7145 | 56.0 | 45864 | 504.3994 | 0.6158 | 0.2365 |
110
+ | 268.4251 | 57.0 | 46683 | 523.2788 | 0.6174 | 0.2346 |
111
+ | 272.3521 | 58.0 | 47502 | 519.0136 | 0.6143 | 0.2350 |
112
+ | 263.5849 | 59.0 | 48321 | 521.5695 | 0.6092 | 0.2333 |
113
+ | 259.7302 | 60.0 | 49140 | 526.7687 | 0.6117 | 0.2332 |
114
+
115
+
116
+ ### Framework versions
117
+
118
+ - Transformers 4.48.3
119
+ - Pytorch 2.5.1
120
+ - Datasets 3.2.0
121
+ - Tokenizers 0.21.0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:333f607d8c5bde383b0df66b941645aff14087ace867691f632270f42fe06511
3
  size 377654192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c39c4abf285d2288f9830e9388bcd24420a9543c7f49681e42701fea8f30b6db
3
  size 377654192