File size: 2,762 Bytes
4814609 40697a3 4814609 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 |
{
"config": {
"model": "LaminiModel",
"model_args": null,
"batch_size": 1,
"batch_sizes": [],
"device": null,
"use_cache": null,
"limit": 100,
"bootstrap_iters": 100000,
"gen_kwargs": null,
"model_dtype": "bfloat16",
"model_name": "icd-ft-mistral-7b-instruct-hf",
"model_sha": "main"
},
"results": {
"product_response_subjective_score": {
"product_response_subjective_score": 3.1875
},
"product_id_precision_score": {
"product_id_precision_score": 0.0
},
"earnings_response_subjective_score": {
"earnings_response_subjective_score": 2.2363636363636363
},
"earnings_precision_score": {
"earnings_precision_score": 0.01818181818181818
},
"icd11_response_subjective_score": {
"icd11_response_subjective_score": 2.2888888888888888
},
"icd11_precision_score": {
"icd11_precision_score": 0.2
},
"mmlu_flan_n_shot_generative_global_facts": {
"exact_match,strict-match": 0.34,
"exact_match_stderr,strict-match": 0.047609522856952344,
"exact_match,flexible-extract": 0.34,
"exact_match_stderr,flexible-extract": 0.047609522856952344,
"alias": "mmlu_flan_n_shot_generative_global_facts"
},
"truthfulqa_gen": {
"bleu_max,none": 17.795881878483,
"bleu_max_stderr,none": 1.813182129677248,
"bleu_acc,none": 0.53,
"bleu_acc_stderr,none": 0.050161355804659205,
"bleu_diff,none": -2.125661702465655,
"bleu_diff_stderr,none": 1.434744252223102,
"rouge1_max,none": 39.008845311741894,
"rouge1_max_stderr,none": 2.114340485165445,
"rouge1_acc,none": 0.52,
"rouge1_acc_stderr,none": 0.050211673156867795,
"rouge1_diff,none": -1.6407676542576224,
"rouge1_diff_stderr,none": 1.797673507015776,
"rouge2_max,none": 27.163127972432466,
"rouge2_max_stderr,none": 2.1194097651441424,
"rouge2_acc,none": 0.49,
"rouge2_acc_stderr,none": 0.05024183937956913,
"rouge2_diff,none": -2.254914453263179,
"rouge2_diff_stderr,none": 1.9148845239446861,
"rougeL_max,none": 36.01824298531865,
"rougeL_max_stderr,none": 2.128032657919697,
"rougeL_acc,none": 0.53,
"rougeL_acc_stderr,none": 0.05016135580465919,
"rougeL_diff,none": -1.8971433636477628,
"rougeL_diff_stderr,none": 1.8261427558039585,
"alias": "truthfulqa_gen"
}
}
} |