| eval/beir-arguana_ndcg@10 = 0.39579 | |
| eval/beir-arguana_recall@100 = 0.94168 | |
| eval/beir-avg_ndcg@10 = 0.342446 | |
| eval/beir-avg_recall@10 = 0.40421200000000007 | |
| eval/beir-avg_recall@100 = 0.624188 | |
| eval/beir-avg_recall@20 = 0.473074 | |
| eval/beir-climate-fever_ndcg@10 = 0.1562 | |
| eval/beir-climate-fever_recall@10 = 0.19275 | |
| eval/beir-climate-fever_recall@100 = 0.42836 | |
| eval/beir-climate-fever_recall@20 = 0.25077 | |
| eval/beir-cqadupstack_ndcg@10 = 0.2689125 | |
| eval/beir-cqadupstack_recall@100 = 0.6019383333333334 | |
| eval/beir-dbpedia-entity_ndcg@10 = 0.29269 | |
| eval/beir-dbpedia-entity_recall@10 = 0.19543 | |
| eval/beir-dbpedia-entity_recall@100 = 0.43193 | |
| eval/beir-dbpedia-entity_recall@20 = 0.26224 | |
| eval/beir-fever_ndcg@10 = 0.55206 | |
| eval/beir-fever_recall@10 = 0.73955 | |
| eval/beir-fever_recall@100 = 0.89197 | |
| eval/beir-fever_recall@20 = 0.79964 | |
| eval/beir-fiqa_ndcg@10 = 0.25367 | |
| eval/beir-fiqa_recall@100 = 0.59399 | |
| eval/beir-hotpotqa_ndcg@10 = 0.49581 | |
| eval/beir-hotpotqa_recall@10 = 0.52519 | |
| eval/beir-hotpotqa_recall@100 = 0.68339 | |
| eval/beir-hotpotqa_recall@20 = 0.57731 | |
| eval/beir-msmarco_ndcg@10 = 0.21547 | |
| eval/beir-msmarco_recall@10 = 0.36814 | |
| eval/beir-msmarco_recall@100 = 0.68529 | |
| eval/beir-msmarco_recall@20 = 0.47541 | |
| eval/beir-nfcorpus_ndcg@10 = 0.29655 | |
| eval/beir-nfcorpus_recall@100 = 0.28415 | |
| eval/beir-nq_ndcg@10 = 0.2742 | |
| eval/beir-nq_recall@100 = 0.79582 | |
| eval/beir-quora_ndcg@10 = 0.77792 | |
| eval/beir-quora_recall@100 = 0.97682 | |
| eval/beir-scidocs_ndcg@10 = 0.15413 | |
| eval/beir-scidocs_recall@100 = 0.37255 | |
| eval/beir-scifact_ndcg@10 = 0.65609 | |
| eval/beir-scifact_recall@100 = 0.91156 | |
| eval/beir-trec-covid_ndcg@10 = 0.56947 | |
| eval/beir-trec-covid_recall@100 = 0.11015 | |
| eval/beir-webis-touche2020_ndcg@10 = 0.18628 | |
| eval/beir-webis-touche2020_recall@100 = 0.4094 | |
| eval/qa-curatedtrec-test-acc@100 = 0.9236311239193083 | |
| eval/qa-curatedtrec-test-acc@20 = 0.8386167146974063 | |
| eval/qa-curatedtrec-test-acc@5 = 0.6455331412103746 | |
| eval/qa-entityqs-macro-acc@100 = 0.7569173874462057 | |
| eval/qa-entityqs-macro-acc@20 = 0.635020152364218 | |
| eval/qa-entityqs-macro-acc@5 = 0.4987727872796979 | |
| eval/qa-nq-test-acc@100 = 0.7894736842105263 | |
| eval/qa-nq-test-acc@20 = 0.6429362880886427 | |
| eval/qa-nq-test-acc@5 = 0.4440443213296399 | |
| eval/qa-squad1-test-acc@100 = 0.7674550614947966 | |
| eval/qa-squad1-test-acc@20 = 0.6085146641438032 | |
| eval/qa-squad1-test-acc@5 = 0.4126773888363292 | |
| eval/qa-trivia-test-acc@100 = 0.8330239547423318 | |
| eval/qa-trivia-test-acc@20 = 0.7475470697427739 | |
| eval/qa-trivia-test-acc@5 = 0.601962344205781 | |
| eval/qa-webq-test-acc@100 = 0.8080708661417323 | |
| eval/qa-webq-test-acc@20 = 0.6712598425196851 | |
| eval/qa-webq-test-acc@5 = 0.46013779527559057 | |
| eval/senteval-CR = 88.1 | |
| eval/senteval-MPQA = 89.04 | |
| eval/senteval-MR = 80.83 | |
| eval/senteval-MRPC = 70.68 | |
| eval/senteval-SICKRelatedness = 0.6952695805138401 | |
| eval/senteval-SST2 = 84.06 | |
| eval/senteval-STS12 = 0.6163833612367604 | |
| eval/senteval-STS13 = 0.7419430794508367 | |
| eval/senteval-STS14 = 0.6677462039090557 | |
| eval/senteval-STS15 = 0.7899640319998164 | |
| eval/senteval-STS16 = 0.7846388532549995 | |
| eval/senteval-STSBenchmark = 0.7564014032837215 | |
| eval/senteval-SUBJ = 95.32 | |
| eval/senteval-TREC = 81.07 | |
| eval/senteval-avg_sts_7 = 0.7217637876641472 | |
| eval/senteval-avg_transfer = 84.15714285714287 | |
| train/global_step = 100000 | |