hf-transformers-bot commited on
Commit
89cc113
·
verified ·
1 Parent(s): 0f4c6ce

Upload 2025-08-26/runs/508-17243184008/ci_results_run_models_gpu/model_results.json with huggingface_hub

Browse files
2025-08-26/runs/508-17243184008/ci_results_run_models_gpu/model_results.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "models_vit": {
3
+ "failed": {
4
+ "PyTorch": {
5
+ "unclassified": 0,
6
+ "single": 0,
7
+ "multi": 0
8
+ },
9
+ "TensorFlow": {
10
+ "unclassified": 0,
11
+ "single": 0,
12
+ "multi": 0
13
+ },
14
+ "Flax": {
15
+ "unclassified": 0,
16
+ "single": 0,
17
+ "multi": 0
18
+ },
19
+ "Tokenizers": {
20
+ "unclassified": 0,
21
+ "single": 0,
22
+ "multi": 0
23
+ },
24
+ "Pipelines": {
25
+ "unclassified": 0,
26
+ "single": 0,
27
+ "multi": 0
28
+ },
29
+ "Trainer": {
30
+ "unclassified": 0,
31
+ "single": 0,
32
+ "multi": 0
33
+ },
34
+ "ONNX": {
35
+ "unclassified": 0,
36
+ "single": 0,
37
+ "multi": 0
38
+ },
39
+ "Auto": {
40
+ "unclassified": 0,
41
+ "single": 0,
42
+ "multi": 0
43
+ },
44
+ "Quantization": {
45
+ "unclassified": 0,
46
+ "single": 0,
47
+ "multi": 0
48
+ },
49
+ "Unclassified": {
50
+ "unclassified": 0,
51
+ "single": 1,
52
+ "multi": 1
53
+ }
54
+ },
55
+ "errors": 0,
56
+ "success": 215,
57
+ "skipped": 207,
58
+ "time_spent": [
59
+ 59.56,
60
+ 59.08
61
+ ],
62
+ "failures": {
63
+ "multi": [
64
+ {
65
+ "line": "tests/models/vit/test_image_processing_vit.py::ViTImageProcessingTest::test_can_compile_fast_image_processor",
66
+ "trace": "(line 744) AssertionError:"
67
+ }
68
+ ],
69
+ "single": [
70
+ {
71
+ "line": "tests/models/vit/test_image_processing_vit.py::ViTImageProcessingTest::test_can_compile_fast_image_processor",
72
+ "trace": "(line 744) AssertionError:"
73
+ }
74
+ ]
75
+ },
76
+ "job_link": {
77
+ "multi": "https://github.com/huggingface/transformers/actions/runs/17243184008/job/48926506693",
78
+ "single": "https://github.com/huggingface/transformers/actions/runs/17243184008/job/48926506651"
79
+ }
80
+ },
81
+ "models_clip": {
82
+ "failed": {
83
+ "PyTorch": {
84
+ "unclassified": 0,
85
+ "single": 1,
86
+ "multi": 1
87
+ },
88
+ "TensorFlow": {
89
+ "unclassified": 0,
90
+ "single": 0,
91
+ "multi": 0
92
+ },
93
+ "Flax": {
94
+ "unclassified": 0,
95
+ "single": 0,
96
+ "multi": 0
97
+ },
98
+ "Tokenizers": {
99
+ "unclassified": 0,
100
+ "single": 0,
101
+ "multi": 0
102
+ },
103
+ "Pipelines": {
104
+ "unclassified": 0,
105
+ "single": 0,
106
+ "multi": 0
107
+ },
108
+ "Trainer": {
109
+ "unclassified": 0,
110
+ "single": 0,
111
+ "multi": 0
112
+ },
113
+ "ONNX": {
114
+ "unclassified": 0,
115
+ "single": 0,
116
+ "multi": 0
117
+ },
118
+ "Auto": {
119
+ "unclassified": 0,
120
+ "single": 0,
121
+ "multi": 0
122
+ },
123
+ "Quantization": {
124
+ "unclassified": 0,
125
+ "single": 0,
126
+ "multi": 0
127
+ },
128
+ "Unclassified": {
129
+ "unclassified": 0,
130
+ "single": 1,
131
+ "multi": 1
132
+ }
133
+ },
134
+ "errors": 0,
135
+ "success": 674,
136
+ "skipped": 942,
137
+ "time_spent": [
138
+ 141.08,
139
+ 143.26
140
+ ],
141
+ "failures": {
142
+ "single": [
143
+ {
144
+ "line": "tests/models/clip/test_image_processing_clip.py::CLIPImageProcessingTest::test_can_compile_fast_image_processor",
145
+ "trace": "(line 744) AssertionError:"
146
+ },
147
+ {
148
+ "line": "tests/models/clip/test_modeling_clip.py::CLIPModelTest::test_config",
149
+ "trace": "(line 553) assert 1 == 2"
150
+ }
151
+ ],
152
+ "multi": [
153
+ {
154
+ "line": "tests/models/clip/test_image_processing_clip.py::CLIPImageProcessingTest::test_can_compile_fast_image_processor",
155
+ "trace": "(line 744) AssertionError:"
156
+ },
157
+ {
158
+ "line": "tests/models/clip/test_modeling_clip.py::CLIPModelTest::test_config",
159
+ "trace": "(line 553) assert 1 == 2"
160
+ }
161
+ ]
162
+ },
163
+ "job_link": {
164
+ "single": "https://github.com/huggingface/transformers/actions/runs/17243184008/job/48926506701",
165
+ "multi": "https://github.com/huggingface/transformers/actions/runs/17243184008/job/48926506687"
166
+ }
167
+ }
168
+ }