Upload folder using huggingface_hub
Browse files- 2_Dense/model.safetensors +1 -1
- README.md +233 -306
- config_sentence_transformers.json +1 -1
- model.safetensors +1 -1
2_Dense/model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 2362528
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:2690907eea4aaaac3e60dd076b650bf7cdd8d9f91783c189fe213b1f202b0e14
|
| 3 |
size 2362528
|
README.md
CHANGED
|
@@ -5,56 +5,64 @@ tags:
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
-
- dataset_size:
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
| 10 |
-
- dataset_size:966
|
| 11 |
base_model: sentence-transformers/LaBSE
|
| 12 |
widget:
|
| 13 |
-
- source_sentence: '
|
| 14 |
-
byed thod [=thams cad] der cing ’bros bar ’gyuro [=’gyur ro]| '
|
| 15 |
sentences:
|
| 16 |
-
-
|
| 17 |
-
- '
|
| 18 |
-
|
| 19 |
-
|
| 20 |
-
|
| 21 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 22 |
sentences:
|
| 23 |
-
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
-
|
| 27 |
-
|
| 28 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 29 |
sentences:
|
| 30 |
-
-
|
| 31 |
-
|
| 32 |
-
|
| 33 |
-
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
-
|
| 39 |
-
-
|
| 40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 41 |
sentences:
|
| 42 |
-
-
|
| 43 |
-
|
| 44 |
-
-
|
| 45 |
-
|
| 46 |
-
ta ka da ya: arhde (=arhade) samyaq sam buddha ya dadya ta: om pünei pünei mahā
|
| 47 |
-
pünei abarimida pünei abari{mi}da pünei ǰnyā na sam bha ro pa či de: om sarvā
|
| 48 |
-
sam sakā ra pa ri šuddhe dharma de ga ga na samud ga de: sva bhā va bišüddhe
|
| 49 |
-
mahā na ya pari vā re sva hā:: '
|
| 50 |
-
- source_sentence: 'dge slong tshul khrims nyos [=nyon mongs] pa’i dus| '
|
| 51 |
sentences:
|
| 52 |
-
-
|
| 53 |
-
-
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
| 57 |
-
ügei-dü idē ögüqčiyin bükün tālaxui xorōdu: '
|
| 58 |
pipeline_tag: sentence-similarity
|
| 59 |
library_name: sentence-transformers
|
| 60 |
---
|
|
@@ -110,9 +118,9 @@ from sentence_transformers import SentenceTransformer
|
|
| 110 |
model = SentenceTransformer("sentence_transformers_model_id")
|
| 111 |
# Run inference
|
| 112 |
sentences = [
|
| 113 |
-
'
|
| 114 |
-
'
|
| 115 |
-
'
|
| 116 |
]
|
| 117 |
embeddings = model.encode(sentences)
|
| 118 |
print(embeddings.shape)
|
|
@@ -166,19 +174,19 @@ You can finetune this model on your own dataset.
|
|
| 166 |
|
| 167 |
#### Unnamed Dataset
|
| 168 |
|
| 169 |
-
* Size:
|
| 170 |
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
|
| 171 |
-
* Approximate statistics based on the first
|
| 172 |
-
| | sentence_0
|
| 173 |
-
|
| 174 |
-
| type | string
|
| 175 |
-
| details | <ul><li>min:
|
| 176 |
* Samples:
|
| 177 |
-
| sentence_0
|
| 178 |
-
|
| 179 |
-
| <code>
|
| 180 |
-
| <code>
|
| 181 |
-
| <code>
|
| 182 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 183 |
```json
|
| 184 |
{
|
|
@@ -192,8 +200,8 @@ You can finetune this model on your own dataset.
|
|
| 192 |
#### Non-Default Hyperparameters
|
| 193 |
|
| 194 |
- `eval_strategy`: steps
|
| 195 |
-
- `per_device_train_batch_size`:
|
| 196 |
-
- `per_device_eval_batch_size`:
|
| 197 |
- `num_train_epochs`: 40
|
| 198 |
- `fp16`: True
|
| 199 |
- `multi_dataset_batch_sampler`: round_robin
|
|
@@ -205,8 +213,8 @@ You can finetune this model on your own dataset.
|
|
| 205 |
- `do_predict`: False
|
| 206 |
- `eval_strategy`: steps
|
| 207 |
- `prediction_loss_only`: True
|
| 208 |
-
- `per_device_train_batch_size`:
|
| 209 |
-
- `per_device_eval_batch_size`:
|
| 210 |
- `per_gpu_train_batch_size`: None
|
| 211 |
- `per_gpu_eval_batch_size`: None
|
| 212 |
- `gradient_accumulation_steps`: 1
|
|
@@ -253,7 +261,7 @@ You can finetune this model on your own dataset.
|
|
| 253 |
- `debug`: []
|
| 254 |
- `dataloader_drop_last`: False
|
| 255 |
- `dataloader_num_workers`: 0
|
| 256 |
-
- `dataloader_prefetch_factor`:
|
| 257 |
- `past_index`: -1
|
| 258 |
- `disable_tqdm`: False
|
| 259 |
- `remove_unused_columns`: True
|
|
@@ -326,249 +334,168 @@ You can finetune this model on your own dataset.
|
|
| 326 |
|
| 327 |
| Epoch | Step |
|
| 328 |
|:-------:|:----:|
|
| 329 |
-
| 0.
|
| 330 |
-
| 0.
|
| 331 |
-
|
|
| 332 |
-
|
|
| 333 |
-
|
|
| 334 |
-
| 1.
|
| 335 |
-
|
|
| 336 |
-
|
|
| 337 |
-
|
|
| 338 |
-
|
|
| 339 |
-
|
|
| 340 |
-
|
|
| 341 |
-
|
|
| 342 |
-
|
|
| 343 |
-
|
|
| 344 |
-
|
|
| 345 |
-
|
|
| 346 |
-
|
|
| 347 |
-
|
|
| 348 |
-
|
|
| 349 |
-
|
|
| 350 |
-
|
|
| 351 |
-
|
|
| 352 |
-
|
|
| 353 |
-
|
|
| 354 |
-
|
|
| 355 |
-
|
|
| 356 |
-
|
|
| 357 |
-
|
|
| 358 |
-
|
|
| 359 |
-
|
|
| 360 |
-
|
|
| 361 |
-
|
|
| 362 |
-
|
|
| 363 |
-
|
|
| 364 |
-
|
|
| 365 |
-
|
|
| 366 |
-
|
|
| 367 |
-
|
|
| 368 |
-
|
|
| 369 |
-
|
|
| 370 |
-
|
|
| 371 |
-
|
|
| 372 |
-
|
|
| 373 |
-
|
|
| 374 |
-
|
|
| 375 |
-
|
|
| 376 |
-
|
|
| 377 |
-
|
|
| 378 |
-
|
|
| 379 |
-
|
|
| 380 |
-
|
|
| 381 |
-
|
|
| 382 |
-
|
|
| 383 |
-
|
|
| 384 |
-
|
|
| 385 |
-
|
|
| 386 |
-
|
|
| 387 |
-
|
|
| 388 |
-
|
|
| 389 |
-
|
|
| 390 |
-
|
|
| 391 |
-
|
|
| 392 |
-
|
|
| 393 |
-
|
|
| 394 |
-
|
|
| 395 |
-
|
|
| 396 |
-
|
|
| 397 |
-
|
|
| 398 |
-
|
|
| 399 |
-
|
|
| 400 |
-
|
|
| 401 |
-
|
|
| 402 |
-
|
|
| 403 |
-
|
|
| 404 |
-
|
|
| 405 |
-
|
|
| 406 |
-
|
|
| 407 |
-
|
|
| 408 |
-
|
|
| 409 |
-
|
|
| 410 |
-
|
|
| 411 |
-
|
|
| 412 |
-
|
|
| 413 |
-
|
|
| 414 |
-
|
|
| 415 |
-
|
|
| 416 |
-
|
|
| 417 |
-
|
|
| 418 |
-
|
|
| 419 |
-
|
|
| 420 |
-
|
|
| 421 |
-
|
|
| 422 |
-
|
|
| 423 |
-
|
|
| 424 |
-
|
|
| 425 |
-
|
|
| 426 |
-
|
|
| 427 |
-
|
|
| 428 |
-
|
|
| 429 |
-
|
|
| 430 |
-
|
|
| 431 |
-
|
|
| 432 |
-
|
|
| 433 |
-
|
|
| 434 |
-
|
|
| 435 |
-
|
|
| 436 |
-
|
|
| 437 |
-
|
|
| 438 |
-
|
|
| 439 |
-
|
|
| 440 |
-
|
|
| 441 |
-
|
|
| 442 |
-
|
|
| 443 |
-
|
|
| 444 |
-
|
|
| 445 |
-
|
|
| 446 |
-
|
|
| 447 |
-
|
|
| 448 |
-
|
|
| 449 |
-
|
|
| 450 |
-
|
|
| 451 |
-
|
|
| 452 |
-
|
|
| 453 |
-
|
|
| 454 |
-
|
|
| 455 |
-
|
|
| 456 |
-
|
|
| 457 |
-
|
|
| 458 |
-
|
|
| 459 |
-
|
|
| 460 |
-
|
|
| 461 |
-
| 0
|
| 462 |
-
|
|
| 463 |
-
|
|
| 464 |
-
|
|
| 465 |
-
|
|
| 466 |
-
|
|
| 467 |
-
|
|
| 468 |
-
|
|
| 469 |
-
|
|
| 470 |
-
|
|
| 471 |
-
|
|
| 472 |
-
|
|
| 473 |
-
|
|
| 474 |
-
|
|
| 475 |
-
|
|
| 476 |
-
|
|
| 477 |
-
|
|
| 478 |
-
|
|
| 479 |
-
|
|
| 480 |
-
|
|
| 481 |
-
|
|
| 482 |
-
|
|
| 483 |
-
|
|
| 484 |
-
|
|
| 485 |
-
|
|
| 486 |
-
|
|
| 487 |
-
|
|
| 488 |
-
|
|
| 489 |
-
|
|
| 490 |
-
|
|
| 491 |
-
| 4.7143 | 99 |
|
| 492 |
-
| 4.8571 | 102 |
|
| 493 |
-
| 5.0 | 105 |
|
| 494 |
-
| 5.1429 | 108 |
|
| 495 |
-
| 5.2857 | 111 |
|
| 496 |
-
| 5.4286 | 114 |
|
| 497 |
-
| 5.5714 | 117 |
|
| 498 |
-
| 5.7143 | 120 |
|
| 499 |
-
| 5.8571 | 123 |
|
| 500 |
-
| 6.0 | 126 |
|
| 501 |
-
| 6.1429 | 129 |
|
| 502 |
-
| 6.2857 | 132 |
|
| 503 |
-
| 6.4286 | 135 |
|
| 504 |
-
| 6.5714 | 138 |
|
| 505 |
-
| 6.7143 | 141 |
|
| 506 |
-
| 6.8571 | 144 |
|
| 507 |
-
| 7.0 | 147 |
|
| 508 |
-
| 7.1429 | 150 |
|
| 509 |
-
| 7.2857 | 153 |
|
| 510 |
-
| 7.4286 | 156 |
|
| 511 |
-
| 7.5714 | 159 |
|
| 512 |
-
| 7.7143 | 162 |
|
| 513 |
-
| 7.8571 | 165 |
|
| 514 |
-
| 8.0 | 168 |
|
| 515 |
-
| 8.1429 | 171 |
|
| 516 |
-
| 8.2857 | 174 |
|
| 517 |
-
| 8.4286 | 177 |
|
| 518 |
-
| 8.5714 | 180 |
|
| 519 |
-
| 8.7143 | 183 |
|
| 520 |
-
| 8.8571 | 186 |
|
| 521 |
-
| 9.0 | 189 |
|
| 522 |
-
| 9.1429 | 192 |
|
| 523 |
-
| 9.2857 | 195 |
|
| 524 |
-
| 9.4286 | 198 |
|
| 525 |
-
| 9.5714 | 201 |
|
| 526 |
-
| 9.7143 | 204 |
|
| 527 |
-
| 9.8571 | 207 |
|
| 528 |
-
| 10.0 | 210 |
|
| 529 |
-
| 10.1429 | 213 |
|
| 530 |
-
| 10.2857 | 216 |
|
| 531 |
-
| 10.4286 | 219 |
|
| 532 |
-
| 10.5714 | 222 |
|
| 533 |
-
| 10.7143 | 225 |
|
| 534 |
-
| 10.8571 | 228 |
|
| 535 |
-
| 11.0 | 231 |
|
| 536 |
-
| 11.1429 | 234 |
|
| 537 |
-
| 11.2857 | 237 |
|
| 538 |
-
| 11.4286 | 240 |
|
| 539 |
-
| 11.5714 | 243 |
|
| 540 |
-
| 11.7143 | 246 |
|
| 541 |
-
| 11.8571 | 249 |
|
| 542 |
-
| 12.0 | 252 |
|
| 543 |
-
| 12.1429 | 255 |
|
| 544 |
-
| 12.2857 | 258 |
|
| 545 |
-
| 12.4286 | 261 |
|
| 546 |
-
| 12.5714 | 264 |
|
| 547 |
-
| 12.7143 | 267 |
|
| 548 |
-
| 12.8571 | 270 |
|
| 549 |
-
| 13.0 | 273 |
|
| 550 |
-
| 13.1429 | 276 |
|
| 551 |
-
| 13.2857 | 279 |
|
| 552 |
-
| 13.4286 | 282 |
|
| 553 |
-
| 13.5714 | 285 |
|
| 554 |
-
| 13.7143 | 288 |
|
| 555 |
-
| 13.8571 | 291 |
|
| 556 |
-
| 14.0 | 294 |
|
| 557 |
-
| 14.1429 | 297 |
|
| 558 |
-
| 14.2857 | 300 |
|
| 559 |
-
| 14.4286 | 303 |
|
| 560 |
-
| 14.5714 | 306 |
|
| 561 |
-
| 14.7143 | 309 |
|
| 562 |
-
| 14.8571 | 312 |
|
| 563 |
-
| 15.0 | 315 |
|
| 564 |
-
| 15.1429 | 318 |
|
| 565 |
-
| 15.2857 | 321 |
|
| 566 |
-
| 15.4286 | 324 |
|
| 567 |
-
| 15.5714 | 327 |
|
| 568 |
-
| 15.7143 | 330 |
|
| 569 |
-
| 15.8571 | 333 |
|
| 570 |
-
| 16.0 | 336 |
|
| 571 |
-
| 16.1429 | 339 |
|
| 572 |
|
| 573 |
</details>
|
| 574 |
|
|
@@ -576,9 +503,9 @@ You can finetune this model on your own dataset.
|
|
| 576 |
- Python: 3.10.0
|
| 577 |
- Sentence Transformers: 5.1.0
|
| 578 |
- Transformers: 4.46.3
|
| 579 |
-
- PyTorch:
|
| 580 |
- Accelerate: 1.1.1
|
| 581 |
-
- Datasets: 4.
|
| 582 |
- Tokenizers: 0.20.3
|
| 583 |
|
| 584 |
## Citation
|
|
|
|
| 5 |
- feature-extraction
|
| 6 |
- dense
|
| 7 |
- generated_from_trainer
|
| 8 |
+
- dataset_size:2099
|
| 9 |
- loss:MultipleNegativesRankingLoss
|
|
|
|
| 10 |
base_model: sentence-transformers/LaBSE
|
| 11 |
widget:
|
| 12 |
+
- source_sentence: 'chos shes bkra shis bkra shis ‘byung . '
|
|
|
|
| 13 |
sentences:
|
| 14 |
+
- nom meden ölzöyitöi ölzöi-tü bolxu .
|
| 15 |
+
- ' om namo bhaγavade abarimida āyur ǰnyā na sübinixči da de čo rā zā ya da ta
|
| 16 |
+
ga da ya arhade samyaq sam buddha ya . daday ta . om pünei pünei mahā pünei
|
| 17 |
+
abarimida pünei abarimida pünei ǰnyā na sam bha ro pa čide . om sarvā sam sakā
|
| 18 |
+
ra pa re šuddhe dharma de ga ga na samud ga de sva bha vā bišüddhe mahā na ya
|
| 19 |
+
pari vā re sva hā . . '
|
| 20 |
+
- kā-ya vā-gī šo-ra . a-ra pa-za-nā ya de na-map . .
|
| 21 |
+
- source_sentence: 'spyan drangs la mchod gnas bya dgos gsungs nas dus de nyid du
|
| 22 |
+
btsun mo dang| blon po dang| phyi ''khor dang| nang ''khor mang pos rgyal po''i
|
| 23 |
+
sngar ''khor nas| '
|
| 24 |
sentences:
|
| 25 |
+
- ere zāni belegeyin tödüi xora orun γadādu dalai toqtobui .
|
| 26 |
+
- ' zalaǰi takiliyin oron bolγoxu kemēn zarliq bolōd . mon daruüda xatun tüšimed
|
| 27 |
+
kigēd . γadādu nököd dotōdü nököd olon-yēr ilete kürēlöülen .'
|
| 28 |
+
- izuurtani köböün töüni šara üsüni [n]ükün buri čü tögönčilen boluqsan suüxü bui
|
| 29 |
+
. suüγād čü izuurtani köböün ene metü erel xangγaxü čindamani erdeniyigi olun
|
| 30 |
+
. üyiledkü olxu boluyu . izuurtani köböün či zurγān züyili tonilγon üyiledüqsen sayin
|
| 31 |
+
sayin kü . izuurtani köböün čini gebelidü orošiqson amitan tedeču xarin ülü ireqči bodhi-sadv
|
| 32 |
+
boluyu . kemēn zarliq bolbui . . kemēgēd ilγon tögösüqsen burxan xamuq töüdker
|
| 33 |
+
teyin arilγaqčidu zarliq bolboi .
|
| 34 |
+
- source_sentence: sangs rgyas 'od srungs la bram ze'i khye'u skar ma'i 'od ces bya
|
| 35 |
+
bas sems bskyed pa dang| sangs rgyas 'od srungs kyis lung bstan pa| bram ze'i
|
| 36 |
+
khye'u khyod tshe lo khri'i dus su zhing khams skar ma'i 'od ces bya bar| sangs
|
| 37 |
+
rgyas mar me mdzad ces bya bar mngon par rdzogs par sangs rgya bar 'gyur ro zhes
|
| 38 |
+
lung bstan nas sangs rgyas so .
|
| 39 |
sentences:
|
| 40 |
+
- gerel sakiqči burxan-du odoni kemēkü birman köböün bodhi sedkil öüskeqsen-dü
|
| 41 |
+
gerel sakiqči burxan eši üzüüleqsen inu . birman köböün či tüme nasulxoi caqtu
|
| 42 |
+
odoni gerel kemēkü tarālanggiyin oron-du dhi-pamka-ra burxan kemēn ilerkei duusun
|
| 43 |
+
burxan bolxu . kemēn eši üzüülēd burxan bolboi .
|
| 44 |
+
- adlidxaxülā dalan yeke tüb niǰēd dusul-ēce bi tōlun čidaxu . saba-yin zokōlliyin
|
| 45 |
+
suduriyin ayimagiyin nigen šülügiyin buyani tōlun ülü čidaxu bui . ada-lidxaxülā
|
| 46 |
+
γangγan möreni xümakiyin tödüi togünčilen boluqsun dayini darün sayitur dou-suqsan
|
| 47 |
+
burxadtu . c[a]q arban <...> xoyor kalab boltolo xubcasun kigēd ebečin-dü šütüküi
|
| 48 |
+
em kigēd . aγoursun-noγoud-yēr sayin kündülel üyiledüqsen buyüni coqcoēce . sabayin
|
| 49 |
+
zokōl-liyin suduriyin ayimagi takin üyileduqsun buyan maši ülemǰi .
|
| 50 |
+
- tende-ēce tenggeriyin e{rke}tu xurmasta beyeyin kemǰē γurba bosxoxoi ayilidxal
|
| 51 |
+
ögȫd . ilγon tögüsün öleqsen ni nayiman nasuni kemǰē . öber dēre tedkiküi ekidu
|
| 52 |
+
xoyor inu xaraqsan-du . eke ögüülebei . nayiman nasuni kemǰē inu . lombiyin
|
| 53 |
+
oi šoγüyin šer<...>yin üzüür-lügē teng bilē kemēn ögüülebei .
|
| 54 |
+
- source_sentence: '’du shes kun gyi don spangs shing . '
|
| 55 |
sentences:
|
| 56 |
+
- aldar ölzöi-töi sayin aldar buyan-tu . .
|
| 57 |
+
- busudiyin ači ülü orkixui yeke yosutai . .
|
| 58 |
+
- xurān medeküi xamugiyin udxa tebčin .
|
| 59 |
+
- source_sentence: 'sangs rgyas ‘byung ba khyod la ‘dud . '
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 60 |
sentences:
|
| 61 |
+
- burxan bolxui čimadu sögödümüi .
|
| 62 |
+
- biraman ögüülebei . bi bükün-dü eldeb zobolong üzeǰi öni bolun xoyino suruqsani
|
| 63 |
+
tula . yeke xān sonosxui durašixü bögǖsü ödüi činēn-yēr bolxu busu .
|
| 64 |
+
- yeke tüšimel tere čü tenggeri-yin šütēni dergede odči eyin kemēn ayiladxabai
|
| 65 |
+
.
|
|
|
|
| 66 |
pipeline_tag: sentence-similarity
|
| 67 |
library_name: sentence-transformers
|
| 68 |
---
|
|
|
|
| 118 |
model = SentenceTransformer("sentence_transformers_model_id")
|
| 119 |
# Run inference
|
| 120 |
sentences = [
|
| 121 |
+
'sangs rgyas ‘byung ba khyod la ‘dud . ',
|
| 122 |
+
'burxan bolxui čimadu sögödümüi .',
|
| 123 |
+
'biraman ögüülebei . bi bükün-dü eldeb zobolong üzeǰi öni bolun xoyino suruqsani tula . yeke xān sonosxui durašixü bögǖsü ödüi činēn-yēr bolxu busu .',
|
| 124 |
]
|
| 125 |
embeddings = model.encode(sentences)
|
| 126 |
print(embeddings.shape)
|
|
|
|
| 174 |
|
| 175 |
#### Unnamed Dataset
|
| 176 |
|
| 177 |
+
* Size: 2,099 training samples
|
| 178 |
* Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code>
|
| 179 |
+
* Approximate statistics based on the first 1000 samples:
|
| 180 |
+
| | sentence_0 | sentence_1 | label |
|
| 181 |
+
|:--------|:------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:--------------------------------------------------------------|
|
| 182 |
+
| type | string | string | float |
|
| 183 |
+
| details | <ul><li>min: 10 tokens</li><li>mean: 64.03 tokens</li><li>max: 256 tokens</li></ul> | <ul><li>min: 7 tokens</li><li>mean: 64.31 tokens</li><li>max: 256 tokens</li></ul> | <ul><li>min: 1.0</li><li>mean: 1.0</li><li>max: 1.0</li></ul> |
|
| 184 |
* Samples:
|
| 185 |
+
| sentence_0 | sentence_1 | label |
|
| 186 |
+
|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-----------------|
|
| 187 |
+
| <code>gdan mthon po gsum bshams pa ’di la\| gcig la ni sangs rgyas bzhugs su gsol cig </code> | <code>γurban önder debisker beledüqsen öüni nigendüni burxani soülγan zalbari </code> | <code>1.0</code> |
|
| 188 |
+
| <code>rku ba spangs nas gzhan la byin pas zas skom 'phel bar 'gyur ro .</code> | <code> xolxoi tebčīd busudtu ökülē idēn undān ürgüǰikü boluyu .</code> | <code>1.0</code> |
|
| 189 |
+
| <code>de'i tshe bcom ldan 'das kyi smin mtshams nas 'od zer mang po bkye nas\| 'od des 'jig rten gyi khams thams cad khyab par byas nas\| mnar med pa'i sems can dmyal ba yan chad la khyab par byas so . sdug bsngal thams cad zhi bar gyur to . de nas slar dbu'i gtsug tor du nub par [(64na)] gyur to .</code> | <code>tere caqtu ilγon tögüsün ülüqseni kümöskü zabsar-ēce olon gerel sacurād . tere gerel-yēr yertünciyin xamoq orudtu tügēn . ayous tamü-ēce inaqši xamuqtu tokiülün üyiledküi . xamuq zobolong maši amurlibai . tegēd xarin ilγon tögüsün üleqsen ni oroyin usnirtu šinggebei .</code> | <code>1.0</code> |
|
| 190 |
* Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters:
|
| 191 |
```json
|
| 192 |
{
|
|
|
|
| 200 |
#### Non-Default Hyperparameters
|
| 201 |
|
| 202 |
- `eval_strategy`: steps
|
| 203 |
+
- `per_device_train_batch_size`: 32
|
| 204 |
+
- `per_device_eval_batch_size`: 32
|
| 205 |
- `num_train_epochs`: 40
|
| 206 |
- `fp16`: True
|
| 207 |
- `multi_dataset_batch_sampler`: round_robin
|
|
|
|
| 213 |
- `do_predict`: False
|
| 214 |
- `eval_strategy`: steps
|
| 215 |
- `prediction_loss_only`: True
|
| 216 |
+
- `per_device_train_batch_size`: 32
|
| 217 |
+
- `per_device_eval_batch_size`: 32
|
| 218 |
- `per_gpu_train_batch_size`: None
|
| 219 |
- `per_gpu_eval_batch_size`: None
|
| 220 |
- `gradient_accumulation_steps`: 1
|
|
|
|
| 261 |
- `debug`: []
|
| 262 |
- `dataloader_drop_last`: False
|
| 263 |
- `dataloader_num_workers`: 0
|
| 264 |
+
- `dataloader_prefetch_factor`: 2
|
| 265 |
- `past_index`: -1
|
| 266 |
- `disable_tqdm`: False
|
| 267 |
- `remove_unused_columns`: True
|
|
|
|
| 334 |
|
| 335 |
| Epoch | Step |
|
| 336 |
|:-------:|:----:|
|
| 337 |
+
| 0.1765 | 3 |
|
| 338 |
+
| 0.3529 | 6 |
|
| 339 |
+
| 0.5294 | 9 |
|
| 340 |
+
| 0.7059 | 12 |
|
| 341 |
+
| 0.8824 | 15 |
|
| 342 |
+
| 1.0 | 17 |
|
| 343 |
+
| 1.0588 | 18 |
|
| 344 |
+
| 1.2353 | 21 |
|
| 345 |
+
| 1.4118 | 24 |
|
| 346 |
+
| 1.5882 | 27 |
|
| 347 |
+
| 1.7647 | 30 |
|
| 348 |
+
| 1.9412 | 33 |
|
| 349 |
+
| 2.0 | 34 |
|
| 350 |
+
| 2.1176 | 36 |
|
| 351 |
+
| 2.2941 | 39 |
|
| 352 |
+
| 2.4706 | 42 |
|
| 353 |
+
| 2.6471 | 45 |
|
| 354 |
+
| 2.8235 | 48 |
|
| 355 |
+
| 3.0 | 51 |
|
| 356 |
+
| 3.1765 | 54 |
|
| 357 |
+
| 3.3529 | 57 |
|
| 358 |
+
| 3.5294 | 60 |
|
| 359 |
+
| 3.7059 | 63 |
|
| 360 |
+
| 3.8824 | 66 |
|
| 361 |
+
| 4.0 | 68 |
|
| 362 |
+
| 4.0588 | 69 |
|
| 363 |
+
| 4.2353 | 72 |
|
| 364 |
+
| 4.4118 | 75 |
|
| 365 |
+
| 4.5882 | 78 |
|
| 366 |
+
| 4.7647 | 81 |
|
| 367 |
+
| 4.9412 | 84 |
|
| 368 |
+
| 5.0 | 85 |
|
| 369 |
+
| 5.1176 | 87 |
|
| 370 |
+
| 5.2941 | 90 |
|
| 371 |
+
| 5.4706 | 93 |
|
| 372 |
+
| 5.6471 | 96 |
|
| 373 |
+
| 5.8235 | 99 |
|
| 374 |
+
| 6.0 | 102 |
|
| 375 |
+
| 6.1765 | 105 |
|
| 376 |
+
| 6.3529 | 108 |
|
| 377 |
+
| 6.5294 | 111 |
|
| 378 |
+
| 6.7059 | 114 |
|
| 379 |
+
| 6.8824 | 117 |
|
| 380 |
+
| 7.0 | 119 |
|
| 381 |
+
| 7.0588 | 120 |
|
| 382 |
+
| 7.2353 | 123 |
|
| 383 |
+
| 7.4118 | 126 |
|
| 384 |
+
| 7.5882 | 129 |
|
| 385 |
+
| 7.7647 | 132 |
|
| 386 |
+
| 7.9412 | 135 |
|
| 387 |
+
| 8.0 | 136 |
|
| 388 |
+
| 8.1176 | 138 |
|
| 389 |
+
| 8.2941 | 141 |
|
| 390 |
+
| 8.4706 | 144 |
|
| 391 |
+
| 8.6471 | 147 |
|
| 392 |
+
| 8.8235 | 150 |
|
| 393 |
+
| 9.0 | 153 |
|
| 394 |
+
| 9.1765 | 156 |
|
| 395 |
+
| 9.3529 | 159 |
|
| 396 |
+
| 9.5294 | 162 |
|
| 397 |
+
| 9.7059 | 165 |
|
| 398 |
+
| 9.8824 | 168 |
|
| 399 |
+
| 10.0 | 170 |
|
| 400 |
+
| 10.0588 | 171 |
|
| 401 |
+
| 10.2353 | 174 |
|
| 402 |
+
| 10.4118 | 177 |
|
| 403 |
+
| 10.5882 | 180 |
|
| 404 |
+
| 10.7647 | 183 |
|
| 405 |
+
| 10.9412 | 186 |
|
| 406 |
+
| 11.0 | 187 |
|
| 407 |
+
| 11.1176 | 189 |
|
| 408 |
+
| 11.2941 | 192 |
|
| 409 |
+
| 11.4706 | 195 |
|
| 410 |
+
| 11.6471 | 198 |
|
| 411 |
+
| 11.8235 | 201 |
|
| 412 |
+
| 12.0 | 204 |
|
| 413 |
+
| 12.1765 | 207 |
|
| 414 |
+
| 12.3529 | 210 |
|
| 415 |
+
| 12.5294 | 213 |
|
| 416 |
+
| 12.7059 | 216 |
|
| 417 |
+
| 12.8824 | 219 |
|
| 418 |
+
| 13.0 | 221 |
|
| 419 |
+
| 13.0588 | 222 |
|
| 420 |
+
| 13.2353 | 225 |
|
| 421 |
+
| 13.4118 | 228 |
|
| 422 |
+
| 13.5882 | 231 |
|
| 423 |
+
| 13.7647 | 234 |
|
| 424 |
+
| 13.9412 | 237 |
|
| 425 |
+
| 14.0 | 238 |
|
| 426 |
+
| 14.1176 | 240 |
|
| 427 |
+
| 14.2941 | 243 |
|
| 428 |
+
| 14.4706 | 246 |
|
| 429 |
+
| 14.6471 | 249 |
|
| 430 |
+
| 14.8235 | 252 |
|
| 431 |
+
| 15.0 | 255 |
|
| 432 |
+
| 15.1765 | 258 |
|
| 433 |
+
| 15.3529 | 261 |
|
| 434 |
+
| 15.5294 | 264 |
|
| 435 |
+
| 15.7059 | 267 |
|
| 436 |
+
| 15.8824 | 270 |
|
| 437 |
+
| 16.0 | 272 |
|
| 438 |
+
| 16.0588 | 273 |
|
| 439 |
+
| 16.2353 | 276 |
|
| 440 |
+
| 16.4118 | 279 |
|
| 441 |
+
| 16.5882 | 282 |
|
| 442 |
+
| 16.7647 | 285 |
|
| 443 |
+
| 16.9412 | 288 |
|
| 444 |
+
| 17.0 | 289 |
|
| 445 |
+
| 17.1176 | 291 |
|
| 446 |
+
| 17.2941 | 294 |
|
| 447 |
+
| 17.4706 | 297 |
|
| 448 |
+
| 17.6471 | 300 |
|
| 449 |
+
| 17.8235 | 303 |
|
| 450 |
+
| 18.0 | 306 |
|
| 451 |
+
| 18.1765 | 309 |
|
| 452 |
+
| 18.3529 | 312 |
|
| 453 |
+
| 18.5294 | 315 |
|
| 454 |
+
| 18.7059 | 318 |
|
| 455 |
+
| 18.8824 | 321 |
|
| 456 |
+
| 19.0 | 323 |
|
| 457 |
+
| 19.0588 | 324 |
|
| 458 |
+
| 19.2353 | 327 |
|
| 459 |
+
| 19.4118 | 330 |
|
| 460 |
+
| 19.5882 | 333 |
|
| 461 |
+
| 19.7647 | 336 |
|
| 462 |
+
| 19.9412 | 339 |
|
| 463 |
+
| 20.0 | 340 |
|
| 464 |
+
| 20.1176 | 342 |
|
| 465 |
+
| 20.2941 | 345 |
|
| 466 |
+
| 20.4706 | 348 |
|
| 467 |
+
| 20.6471 | 351 |
|
| 468 |
+
| 20.8235 | 354 |
|
| 469 |
+
| 21.0 | 357 |
|
| 470 |
+
| 21.1765 | 360 |
|
| 471 |
+
| 21.3529 | 363 |
|
| 472 |
+
| 21.5294 | 366 |
|
| 473 |
+
| 21.7059 | 369 |
|
| 474 |
+
| 21.8824 | 372 |
|
| 475 |
+
| 22.0 | 374 |
|
| 476 |
+
| 22.0588 | 375 |
|
| 477 |
+
| 22.2353 | 378 |
|
| 478 |
+
| 22.4118 | 381 |
|
| 479 |
+
| 22.5882 | 384 |
|
| 480 |
+
| 22.7647 | 387 |
|
| 481 |
+
| 22.9412 | 390 |
|
| 482 |
+
| 23.0 | 391 |
|
| 483 |
+
| 23.1176 | 393 |
|
| 484 |
+
| 23.2941 | 396 |
|
| 485 |
+
| 23.4706 | 399 |
|
| 486 |
+
| 23.6471 | 402 |
|
| 487 |
+
| 23.8235 | 405 |
|
| 488 |
+
| 24.0 | 408 |
|
| 489 |
+
| 24.1765 | 411 |
|
| 490 |
+
| 24.3529 | 414 |
|
| 491 |
+
| 24.5294 | 417 |
|
| 492 |
+
| 24.7059 | 420 |
|
| 493 |
+
| 24.8824 | 423 |
|
| 494 |
+
| 25.0 | 425 |
|
| 495 |
+
| 25.0588 | 426 |
|
| 496 |
+
| 25.2353 | 429 |
|
| 497 |
+
| 25.4118 | 432 |
|
| 498 |
+
| 25.5882 | 435 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 499 |
|
| 500 |
</details>
|
| 501 |
|
|
|
|
| 503 |
- Python: 3.10.0
|
| 504 |
- Sentence Transformers: 5.1.0
|
| 505 |
- Transformers: 4.46.3
|
| 506 |
+
- PyTorch: 1.12.1+cu113
|
| 507 |
- Accelerate: 1.1.1
|
| 508 |
+
- Datasets: 4.4.1
|
| 509 |
- Tokenizers: 0.20.3
|
| 510 |
|
| 511 |
## Citation
|
config_sentence_transformers.json
CHANGED
|
@@ -2,7 +2,7 @@
|
|
| 2 |
"__version__": {
|
| 3 |
"sentence_transformers": "5.1.0",
|
| 4 |
"transformers": "4.46.3",
|
| 5 |
-
"pytorch": "
|
| 6 |
},
|
| 7 |
"model_type": "SentenceTransformer",
|
| 8 |
"prompts": {
|
|
|
|
| 2 |
"__version__": {
|
| 3 |
"sentence_transformers": "5.1.0",
|
| 4 |
"transformers": "4.46.3",
|
| 5 |
+
"pytorch": "1.12.1+cu113"
|
| 6 |
},
|
| 7 |
"model_type": "SentenceTransformer",
|
| 8 |
"prompts": {
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 1883730160
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23b6e8e2461f0e8528f77126ef5b3df5f1b3b5d00c2f788e233da28fb5d70aa8
|
| 3 |
size 1883730160
|