hf-transformers-bot commited on
Commit
465489f
·
verified ·
1 Parent(s): f8532e7

Upload 2025-09-03/runs/4997-17438034315/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-09-03/runs/4997-17438034315/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_qwen2_5_omni": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 0
8
+ },
9
+ "TensorFlow": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Flax": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Tokenizers": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "Pipelines": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Trainer": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "ONNX": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Auto": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ },
44
+ "Quantization": {
45
+ "unclassified": 0,
46
+ "single": 0,
47
+ "multi": 0
48
+ },
49
+ "Unclassified": {
50
+ "unclassified": 0,
51
+ "single": 2,
52
+ "multi": 2
53
+ }
54
+ },
55
+ "errors": 0,
56
+ "success": 309,
57
+ "skipped": 127,
58
+ "time_spent": [
59
+ 193.19,
60
+ 198.33
61
+ ],
62
+ "failures": {
63
+ "multi": [
64
+ {
65
+ "line": "tests/models/qwen2_5_omni/test_processing_qwen2_5_omni.py::Qwen2_5OmniProcessorTest::test_apply_chat_template_audio_2",
66
+ "trace": "(line 829) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
67
+ },
68
+ {
69
+ "line": "tests/models/qwen2_5_omni/test_processing_qwen2_5_omni.py::Qwen2_5OmniProcessorTest::test_apply_chat_template_audio_3",
70
+ "trace": "(line 829) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
71
+ }
72
+ ],
73
+ "single": [
74
+ {
75
+ "line": "tests/models/qwen2_5_omni/test_processing_qwen2_5_omni.py::Qwen2_5OmniProcessorTest::test_apply_chat_template_audio_2",
76
+ "trace": "(line 829) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
77
+ },
78
+ {
79
+ "line": "tests/models/qwen2_5_omni/test_processing_qwen2_5_omni.py::Qwen2_5OmniProcessorTest::test_apply_chat_template_audio_3",
80
+ "trace": "(line 829) RuntimeError: The frame has 0 channels, expected 1. If you are hitting this, it may be because you are using a buggy FFmpeg version. FFmpeg4 is known to fail here in some valid scenarios. Try to upgrade FFmpeg?"
81
+ }
82
+ ]
83
+ },
84
+ "job_link": {
85
+ "multi": "https://github.com/huggingface/transformers/actions/runs/17438034315/job/49513900369",
86
+ "single": "https://github.com/huggingface/transformers/actions/runs/17438034315/job/49513900320"
87
+ }
88
+ },
89
+ "models_qwen2_5_vl": {
90
+ "failed": {
91
+ "PyTorch": {
92
+ "unclassified": 0,
93
+ "single": 2,
94
+ "multi": 2
95
+ },
96
+ "TensorFlow": {
97
+ "unclassified": 0,
98
+ "single": 0,
99
+ "multi": 0
100
+ },
101
+ "Flax": {
102
+ "unclassified": 0,
103
+ "single": 0,
104
+ "multi": 0
105
+ },
106
+ "Tokenizers": {
107
+ "unclassified": 0,
108
+ "single": 0,
109
+ "multi": 0
110
+ },
111
+ "Pipelines": {
112
+ "unclassified": 0,
113
+ "single": 0,
114
+ "multi": 0
115
+ },
116
+ "Trainer": {
117
+ "unclassified": 0,
118
+ "single": 0,
119
+ "multi": 0
120
+ },
121
+ "ONNX": {
122
+ "unclassified": 0,
123
+ "single": 0,
124
+ "multi": 0
125
+ },
126
+ "Auto": {
127
+ "unclassified": 0,
128
+ "single": 0,
129
+ "multi": 0
130
+ },
131
+ "Quantization": {
132
+ "unclassified": 0,
133
+ "single": 0,
134
+ "multi": 0
135
+ },
136
+ "Unclassified": {
137
+ "unclassified": 0,
138
+ "single": 0,
139
+ "multi": 0
140
+ }
141
+ },
142
+ "errors": 0,
143
+ "success": 321,
144
+ "skipped": 123,
145
+ "time_spent": [
146
+ 235.28,
147
+ 231.48
148
+ ],
149
+ "failures": {
150
+ "multi": [
151
+ {
152
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
153
+ "trace": "(line 675) AssertionError: 'syst[73 chars]ant\\n addCriterion\\nThe dog in the picture app[95 chars]h is' != 'syst[73 chars]ant\\nThe dog in the picture appears to be a La[94 chars]t in'"
154
+ },
155
+ {
156
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
157
+ "trace": "(line 675) AssertionError: Lists differ: ['sys[216 chars]in', 'system\\nYou are a helpful assistant.\\nus[29 chars]aks'] != ['sys[216 chars]in', \"system\\nYou are a helpful assistant.\\nus[162 chars]ing\"]"
158
+ }
159
+ ],
160
+ "single": [
161
+ {
162
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_different_resolutions",
163
+ "trace": "(line 675) AssertionError: 'syst[73 chars]ant\\n addCriterion\\nThe dog in the picture app[95 chars]h is' != 'syst[73 chars]ant\\nThe dog in the picture appears to be a La[94 chars]t in'"
164
+ },
165
+ {
166
+ "line": "tests/models/qwen2_5_vl/test_modeling_qwen2_5_vl.py::Qwen2_5_VLIntegrationTest::test_small_model_integration_test_batch_wo_image_flashatt2",
167
+ "trace": "(line 675) AssertionError: Lists differ: ['sys[216 chars]in', 'system\\nYou are a helpful assistant.\\nus[29 chars]aks'] != ['sys[216 chars]in', \"system\\nYou are a helpful assistant.\\nus[162 chars]ing\"]"
168
+ }
169
+ ]
170
+ },
171
+ "job_link": {
172
+ "multi": "https://github.com/huggingface/transformers/actions/runs/17438034315/job/49513900372",
173
+ "single": "https://github.com/huggingface/transformers/actions/runs/17438034315/job/49513900392"
174
+ }
175
+ }
176
+ }