Bekhouche commited on
Commit
733ec04
·
verified ·
1 Parent(s): b02caf5

Upload Pipeline

Browse files
Files changed (2) hide show
  1. config.json +209 -209
  2. model.safetensors +2 -2
config.json CHANGED
@@ -1,209 +1,209 @@
1
- {
2
- "inputs": [
3
- "images"
4
- ],
5
- "modules": {
6
- "avg_pool": {
7
- "config": {
8
- "args": {
9
- "output_size": [
10
- null,
11
- 1
12
- ]
13
- }
14
- },
15
- "type": "torch.nn.AdaptiveAvgPool2d"
16
- },
17
- "feature_extraction": {
18
- "config": {
19
- "args": {
20
- "input_channel": 1,
21
- "output_channel": 512,
22
- "variant": "DTRB"
23
- }
24
- },
25
- "type": "DeepTextRecognition.VGGModel"
26
- },
27
- "permute": {
28
- "config": {
29
- "args": {
30
- "dims": [
31
- 0,
32
- 3,
33
- 1,
34
- 2
35
- ]
36
- }
37
- },
38
- "type": "torch.permute"
39
- },
40
- "prediction": {
41
- "config": {
42
- "args": {
43
- "in_features": 256,
44
- "out_features": 37
45
- }
46
- },
47
- "type": "torch.nn.Linear"
48
- },
49
- "processing": {
50
- "config": {
51
- "args": {
52
- "channels_size": 1,
53
- "image_size": [
54
- 32,
55
- 100
56
- ],
57
- "padding": "left"
58
- }
59
- },
60
- "type": "DeepTextRecognition.ImageProcessor"
61
- },
62
- "sequence_modeling": {
63
- "config": {
64
- "args": {
65
- "hidden_sizes": [
66
- 256,
67
- 256
68
- ],
69
- "input_size": 512,
70
- "output_size": 256
71
- }
72
- },
73
- "type": "DeepTextRecognition.BiLSTMModel"
74
- },
75
- "squeeze": {
76
- "config": {
77
- "args": {
78
- "dim": 3
79
- }
80
- },
81
- "type": "torch.squeeze"
82
- },
83
- "tokenizer": {
84
- "config": {
85
- "args": {
86
- "characters": [
87
- "0",
88
- "1",
89
- "2",
90
- "3",
91
- "4",
92
- "5",
93
- "6",
94
- "7",
95
- "8",
96
- "9",
97
- "a",
98
- "b",
99
- "c",
100
- "d",
101
- "e",
102
- "f",
103
- "g",
104
- "h",
105
- "i",
106
- "j",
107
- "k",
108
- "l",
109
- "m",
110
- "n",
111
- "o",
112
- "p",
113
- "q",
114
- "r",
115
- "s",
116
- "t",
117
- "u",
118
- "v",
119
- "w",
120
- "x",
121
- "y",
122
- "z"
123
- ],
124
- "max_length": 25
125
- }
126
- },
127
- "type": "DeepTextRecognition.CTCTokenizer"
128
- }
129
- },
130
- "order": [
131
- "processing",
132
- "feature_extraction",
133
- "permute",
134
- "avg_pool",
135
- "squeeze",
136
- "sequence_modeling",
137
- "prediction",
138
- "tokenizer"
139
- ],
140
- "outputs": [
141
- "tokenizer:labels"
142
- ],
143
- "routing": {
144
- "avg_pool": {
145
- "inputs": [
146
- "permute:permuted_features"
147
- ],
148
- "outputs": [
149
- "avg_pool:pooled_features"
150
- ]
151
- },
152
- "feature_extraction": {
153
- "inputs": [
154
- "processing:processed_images"
155
- ],
156
- "outputs": [
157
- "feature_extraction:extracted_features"
158
- ]
159
- },
160
- "permute": {
161
- "inputs": [
162
- "feature_extraction:extracted_features"
163
- ],
164
- "outputs": [
165
- "permute:permuted_features"
166
- ]
167
- },
168
- "prediction": {
169
- "inputs": [
170
- "sequence_modeling:modeled_features"
171
- ],
172
- "outputs": [
173
- "prediction:predictions"
174
- ]
175
- },
176
- "processing": {
177
- "inputs": [
178
- "images"
179
- ],
180
- "outputs": [
181
- "processing:processed_images"
182
- ]
183
- },
184
- "sequence_modeling": {
185
- "inputs": [
186
- "squeeze:squeezed_features"
187
- ],
188
- "outputs": [
189
- "sequence_modeling:modeled_features"
190
- ]
191
- },
192
- "squeeze": {
193
- "inputs": [
194
- "avg_pool:pooled_features"
195
- ],
196
- "outputs": [
197
- "squeeze:squeezed_features"
198
- ]
199
- },
200
- "tokenizer": {
201
- "inputs": [
202
- "prediction:predictions"
203
- ],
204
- "outputs": [
205
- "tokenizer:labels"
206
- ]
207
- }
208
- }
209
- }
 
1
+ {
2
+ "inputs": [
3
+ "images"
4
+ ],
5
+ "modules": {
6
+ "avg_pool": {
7
+ "config": {
8
+ "args": {
9
+ "output_size": [
10
+ null,
11
+ 1
12
+ ]
13
+ }
14
+ },
15
+ "type": "DeepTextRecognition.AdaptiveAvgPoolModule"
16
+ },
17
+ "feature_extraction": {
18
+ "config": {
19
+ "args": {
20
+ "input_channel": 1,
21
+ "output_channel": 512,
22
+ "variant": "DTRB"
23
+ }
24
+ },
25
+ "type": "DeepTextRecognition.VGGModel"
26
+ },
27
+ "permute": {
28
+ "config": {
29
+ "args": {
30
+ "dims": [
31
+ 0,
32
+ 3,
33
+ 1,
34
+ 2
35
+ ]
36
+ }
37
+ },
38
+ "type": "DeepTextRecognition.PermuteModule"
39
+ },
40
+ "prediction": {
41
+ "config": {
42
+ "args": {
43
+ "in_features": 256,
44
+ "out_features": 37
45
+ }
46
+ },
47
+ "type": "torch.nn.Linear"
48
+ },
49
+ "processing": {
50
+ "config": {
51
+ "args": {
52
+ "channels_size": 1,
53
+ "image_size": [
54
+ 32,
55
+ 100
56
+ ],
57
+ "padding": "left"
58
+ }
59
+ },
60
+ "type": "DeepTextRecognition.ImageProcessor"
61
+ },
62
+ "sequence_modeling": {
63
+ "config": {
64
+ "args": {
65
+ "hidden_sizes": [
66
+ 256,
67
+ 256
68
+ ],
69
+ "input_size": 512,
70
+ "output_size": 256
71
+ }
72
+ },
73
+ "type": "DeepTextRecognition.BiLSTMModel"
74
+ },
75
+ "squeeze": {
76
+ "config": {
77
+ "args": {
78
+ "dim": 3
79
+ }
80
+ },
81
+ "type": "DeepTextRecognition.SqueezeModule"
82
+ },
83
+ "tokenizer": {
84
+ "config": {
85
+ "args": {
86
+ "characters": [
87
+ "0",
88
+ "1",
89
+ "2",
90
+ "3",
91
+ "4",
92
+ "5",
93
+ "6",
94
+ "7",
95
+ "8",
96
+ "9",
97
+ "a",
98
+ "b",
99
+ "c",
100
+ "d",
101
+ "e",
102
+ "f",
103
+ "g",
104
+ "h",
105
+ "i",
106
+ "j",
107
+ "k",
108
+ "l",
109
+ "m",
110
+ "n",
111
+ "o",
112
+ "p",
113
+ "q",
114
+ "r",
115
+ "s",
116
+ "t",
117
+ "u",
118
+ "v",
119
+ "w",
120
+ "x",
121
+ "y",
122
+ "z"
123
+ ],
124
+ "max_length": 25
125
+ }
126
+ },
127
+ "type": "DeepTextRecognition.CTCTokenizer"
128
+ }
129
+ },
130
+ "order": [
131
+ "processing",
132
+ "feature_extraction",
133
+ "permute",
134
+ "avg_pool",
135
+ "squeeze",
136
+ "sequence_modeling",
137
+ "prediction",
138
+ "tokenizer"
139
+ ],
140
+ "outputs": [
141
+ "tokenizer:labels"
142
+ ],
143
+ "routing": {
144
+ "avg_pool": {
145
+ "inputs": [
146
+ "permute:permuted_features"
147
+ ],
148
+ "outputs": [
149
+ "avg_pool:pooled_features"
150
+ ]
151
+ },
152
+ "feature_extraction": {
153
+ "inputs": [
154
+ "processing:processed_images"
155
+ ],
156
+ "outputs": [
157
+ "feature_extraction:extracted_features"
158
+ ]
159
+ },
160
+ "permute": {
161
+ "inputs": [
162
+ "feature_extraction:extracted_features"
163
+ ],
164
+ "outputs": [
165
+ "permute:permuted_features"
166
+ ]
167
+ },
168
+ "prediction": {
169
+ "inputs": [
170
+ "sequence_modeling:modeled_features"
171
+ ],
172
+ "outputs": [
173
+ "prediction:predictions"
174
+ ]
175
+ },
176
+ "processing": {
177
+ "inputs": [
178
+ "images"
179
+ ],
180
+ "outputs": [
181
+ "processing:processed_images"
182
+ ]
183
+ },
184
+ "sequence_modeling": {
185
+ "inputs": [
186
+ "squeeze:squeezed_features"
187
+ ],
188
+ "outputs": [
189
+ "sequence_modeling:modeled_features"
190
+ ]
191
+ },
192
+ "squeeze": {
193
+ "inputs": [
194
+ "avg_pool:pooled_features"
195
+ ],
196
+ "outputs": [
197
+ "squeeze:squeezed_features"
198
+ ]
199
+ },
200
+ "tokenizer": {
201
+ "inputs": [
202
+ "prediction:predictions"
203
+ ],
204
+ "outputs": [
205
+ "tokenizer:labels"
206
+ ]
207
+ }
208
+ }
209
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:28a5829dec46739ec8067ee9659eb258b021e7b6bc1e477ebfa82fddf885604b
3
- size 33819884
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:50ad3517c9804ca396ec703884df841f6b281836b0ffec83a52832cf133c67b6
3
+ size 33820236