update README.md
Browse files
README.md
CHANGED
|
@@ -71,7 +71,7 @@ We evaluate our model on several academic benchmarks then compare with other sim
|
|
| 71 |
| HellaSwag(0-shot) | 82.03 | 81.57 | 83.32 |
|
| 72 |
|
| 73 |
|
| 74 |
-
**Note:** To facilitate reproduction, the results of common benchmarks are generated by [OpenCompass](https://github.com/open-compass/opencompass) except humaneval and mbpp as we experience code timeout and postprocess issues.
|
| 75 |
|
| 76 |
### Chat Model
|
| 77 |
|
|
@@ -85,9 +85,6 @@ We present the performance results of our chat model and other LLM on various st
|
|
| 85 |
| Arena-Hard | 24.2 | 42.6 | 43.1 |
|
| 86 |
| GSM8K | 81.42 | 79.45 | 84.04 |
|
| 87 |
| MATH | 42.28 | 54.06 | 51.48 |
|
| 88 |
-
| USMLE | 58.70 | 55.84 | 79.70 |
|
| 89 |
-
| CFA 2.0 | 35.5 | 42.5 | 62.75 |
|
| 90 |
-
|
| 91 |
|
| 92 |
### Long Context
|
| 93 |
|
|
|
|
| 71 |
| HellaSwag(0-shot) | 82.03 | 81.57 | 83.32 |
|
| 72 |
|
| 73 |
|
| 74 |
+
**Note:** To facilitate reproduction, the results of common benchmarks are generated by [OpenCompass](https://github.com/open-compass/opencompass) except humaneval and mbpp as we experience code timeout and postprocess issues.
|
| 75 |
|
| 76 |
### Chat Model
|
| 77 |
|
|
|
|
| 85 |
| Arena-Hard | 24.2 | 42.6 | 43.1 |
|
| 86 |
| GSM8K | 81.42 | 79.45 | 84.04 |
|
| 87 |
| MATH | 42.28 | 54.06 | 51.48 |
|
|
|
|
|
|
|
|
|
|
| 88 |
|
| 89 |
### Long Context
|
| 90 |
|