johnpotter commited on
Commit
67f847a
·
verified ·
1 Parent(s): 53d902c

fixed typos to README

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -76,8 +76,8 @@ https://portal.nousresearch.com/
76
 
77
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/O_sgWq4CVPuxuKYqHWkkN.png)
78
 
79
- *Reasoning ON benchmarks aquired by running HuggingFace's open-r1 reasoning mode evaluation suite, and scores for reasoning mode OFF aquired by running LM-Eval-Harness Benchmark Suite*
80
- *Upper bound determined by measuring the % gained over Hermes 3 3 & 70b by MATH_VERIFY compared to eleuther eval harness, which ranged betweeen 33% and 50% gain in MATH Hard benchmark on retested models by them compared to eval harness reported scores*
81
 
82
  ## Benchmarks in **Non-Reasoning Mode** against Llama-3.1-8B-Instruct
83
 
 
76
 
77
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6317aade83d8d2fd903192d9/O_sgWq4CVPuxuKYqHWkkN.png)
78
 
79
+ *Reasoning ON benchmarks acquired by running HuggingFace's open-r1 reasoning mode evaluation suite, and scores for reasoning mode OFF acquired by running LM-Eval-Harness Benchmark Suite*
80
+ *Upper bound determined by measuring the % gained over Hermes 3 3 & 70b by MATH_VERIFY compared to eleuther eval harness, which ranged between 33% and 50% gain in MATH Hard benchmark on retested models by them compared to eval harness reported scores*
81
 
82
  ## Benchmarks in **Non-Reasoning Mode** against Llama-3.1-8B-Instruct
83