codefuse-admin
commited on
Commit
·
ae72fb2
1
Parent(s):
60aa158
add json file
Browse files- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/codefuse-ai_CodeFuse-DeepSeek-33b_codefuse-admin.json +1 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_humaneval_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-cpp_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-d_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-java_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-jl_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-js_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-lua_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-php_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-r_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-rkt_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-rs_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-swift_CodeFuse-DeepSeek-33b.json +0 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_humaneval_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-cpp_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-d_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-java_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-jl_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-js_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-lua_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-php_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-r_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-rkt_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-rs_CodeFuse-DeepSeek-33b.json +44 -0
- community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-swift_CodeFuse-DeepSeek-33b.json +44 -0
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/codefuse-ai_CodeFuse-DeepSeek-33b_codefuse-admin.json
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"results": [{"task": "multiple-r", "pass@1": 0.37888198757763975}, {"task": "humaneval", "pass@1": 0.7682926829268293}, {"task": "multiple-lua", "pass@1": 0.4720496894409938}, {"task": "multiple-php", "pass@1": 0.577639751552795}, {"task": "multiple-d", "pass@1": 0.2564102564102564}, {"task": "multiple-jl", "pass@1": 0.4528301886792453}, {"task": "multiple-cpp", "pass@1": 0.6521739130434783}, {"task": "multiple-java", "pass@1": 0.5822784810126582}, {"task": "multiple-rs", "pass@1": 0.5448717948717948}, {"task": "multiple-swift", "pass@1": 0.5093167701863354}, {"task": "multiple-js", "pass@1": 0.6521739130434783}, {"task": "multiple-rkt", "pass@1": 0.33540372670807456}], "meta": {"model": "codefuse-ai/CodeFuse-DeepSeek-33b"}}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_humaneval_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-cpp_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-d_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-java_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-jl_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-js_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-lua_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-php_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-r_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-rkt_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-rs_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/generations_CodeFuse-DeepSeek-33b/generations_multiple-swift_CodeFuse-DeepSeek-33b.json
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_humaneval_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"humaneval": {
|
| 3 |
+
"pass@1": 0.7682926829268293
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "humaneval",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_humaneval_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_humaneval_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-cpp_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-cpp": {
|
| 3 |
+
"pass@1": 0.6521739130434783
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-cpp",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-cpp_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-cpp_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-d_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-d": {
|
| 3 |
+
"pass@1": 0.2564102564102564
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-d",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-d_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-d_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-java_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-java": {
|
| 3 |
+
"pass@1": 0.5822784810126582
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-java",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-java_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-java_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-jl_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-jl": {
|
| 3 |
+
"pass@1": 0.4528301886792453
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-jl",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-jl_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-jl_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-js_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-js": {
|
| 3 |
+
"pass@1": 0.6521739130434783
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-js",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-js_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-js_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-lua_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-lua": {
|
| 3 |
+
"pass@1": 0.4720496894409938
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-lua",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-lua_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-lua_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-php_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-php": {
|
| 3 |
+
"pass@1": 0.577639751552795
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-php",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-php_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-php_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-r_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-r": {
|
| 3 |
+
"pass@1": 0.37888198757763975
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-r",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-r_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-r_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-rkt_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-rkt": {
|
| 3 |
+
"pass@1": 0.33540372670807456
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-rkt",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-rkt_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-rkt_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-rs_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-rs": {
|
| 3 |
+
"pass@1": 0.5448717948717948
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-rs",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-rs_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-rs_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|
community_results/codefuse-ai_codefuse-deepseek-33b_codefuse-admin/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-swift_CodeFuse-DeepSeek-33b.json
ADDED
|
@@ -0,0 +1,44 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"multiple-swift": {
|
| 3 |
+
"pass@1": 0.5093167701863354
|
| 4 |
+
},
|
| 5 |
+
"config": {
|
| 6 |
+
"prefix": "",
|
| 7 |
+
"suffix": "",
|
| 8 |
+
"add_special_tokens": true,
|
| 9 |
+
"do_sample": false,
|
| 10 |
+
"temperature": 0.2,
|
| 11 |
+
"top_k": 0,
|
| 12 |
+
"top_p": 0.95,
|
| 13 |
+
"n_samples": 1,
|
| 14 |
+
"eos": "<|im_end|>",
|
| 15 |
+
"seed": 999999999,
|
| 16 |
+
"model": "codefuse-ai/CodeFuse-DeepSeek-33b",
|
| 17 |
+
"modeltype": "causal",
|
| 18 |
+
"peft_model": null,
|
| 19 |
+
"revision": null,
|
| 20 |
+
"use_auth_token": true,
|
| 21 |
+
"trust_remote_code": true,
|
| 22 |
+
"tasks": "multiple-swift",
|
| 23 |
+
"instruction_tokens": null,
|
| 24 |
+
"batch_size": 1,
|
| 25 |
+
"max_length_generation": 1024,
|
| 26 |
+
"precision": "bf16",
|
| 27 |
+
"load_in_8bit": false,
|
| 28 |
+
"load_in_4bit": false,
|
| 29 |
+
"limit": null,
|
| 30 |
+
"limit_start": 0,
|
| 31 |
+
"postprocess": true,
|
| 32 |
+
"allow_code_execution": true,
|
| 33 |
+
"generation_only": false,
|
| 34 |
+
"load_generations_path": "/app/generations_multiple-swift_CodeFuse-DeepSeek-33b.json",
|
| 35 |
+
"load_data_path": null,
|
| 36 |
+
"metric_output_path": "/app/metrics_CodeFuse-DeepSeek-33b/metrics_multiple-swift_CodeFuse-DeepSeek-33b.json",
|
| 37 |
+
"save_generations": false,
|
| 38 |
+
"save_generations_path": "generations.json",
|
| 39 |
+
"save_references": false,
|
| 40 |
+
"prompt": "prompt",
|
| 41 |
+
"max_memory_per_gpu": null,
|
| 42 |
+
"check_references": false
|
| 43 |
+
}
|
| 44 |
+
}
|