marcbarouky commited on
Commit
056711b
·
verified ·
1 Parent(s): 40d8de7

Upload folder using huggingface_hub

Browse files
depth_estimator/config.json ADDED
@@ -0,0 +1,472 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": "d49bafe6f381b0fe37ccfc4c8f6a23424b09d6ef",
3
+ "_name_or_path": "/home/techoffice/.cache/huggingface/hub/models--stabilityai--stable-diffusion-2-depth/snapshots/d49bafe6f381b0fe37ccfc4c8f6a23424b09d6ef/depth_estimator",
4
+ "add_projection": false,
5
+ "architectures": [
6
+ "DPTForDepthEstimation"
7
+ ],
8
+ "attention_probs_dropout_prob": 0.0,
9
+ "auxiliary_loss_weight": 0.4,
10
+ "backbone": null,
11
+ "backbone_config": {
12
+ "_name_or_path": "",
13
+ "add_cross_attention": false,
14
+ "architectures": null,
15
+ "bad_words_ids": null,
16
+ "begin_suppress_tokens": null,
17
+ "bos_token_id": null,
18
+ "chunk_size_feed_forward": 0,
19
+ "cross_attention_hidden_size": null,
20
+ "decoder_start_token_id": null,
21
+ "depths": [
22
+ 3,
23
+ 4,
24
+ 9
25
+ ],
26
+ "diversity_penalty": 0.0,
27
+ "do_sample": false,
28
+ "drop_path_rate": 0.0,
29
+ "early_stopping": false,
30
+ "embedding_dynamic_padding": true,
31
+ "embedding_size": 64,
32
+ "encoder_no_repeat_ngram_size": 0,
33
+ "eos_token_id": null,
34
+ "exponential_decay_length_penalty": null,
35
+ "finetuning_task": null,
36
+ "forced_bos_token_id": null,
37
+ "forced_eos_token_id": null,
38
+ "global_padding": "SAME",
39
+ "hidden_act": "relu",
40
+ "hidden_sizes": [
41
+ 256,
42
+ 512,
43
+ 1024,
44
+ 2048
45
+ ],
46
+ "id2label": {
47
+ "0": "LABEL_0",
48
+ "1": "LABEL_1"
49
+ },
50
+ "is_decoder": false,
51
+ "is_encoder_decoder": false,
52
+ "label2id": {
53
+ "LABEL_0": 0,
54
+ "LABEL_1": 1
55
+ },
56
+ "layer_type": "bottleneck",
57
+ "length_penalty": 1.0,
58
+ "max_length": 20,
59
+ "min_length": 0,
60
+ "model_type": "bit",
61
+ "no_repeat_ngram_size": 0,
62
+ "num_beam_groups": 1,
63
+ "num_beams": 1,
64
+ "num_channels": 3,
65
+ "num_groups": 32,
66
+ "num_return_sequences": 1,
67
+ "out_features": [
68
+ "stage1",
69
+ "stage2",
70
+ "stage3"
71
+ ],
72
+ "out_indices": [
73
+ 1,
74
+ 2,
75
+ 3
76
+ ],
77
+ "output_attentions": false,
78
+ "output_hidden_states": false,
79
+ "output_scores": false,
80
+ "output_stride": 32,
81
+ "pad_token_id": null,
82
+ "prefix": null,
83
+ "problem_type": null,
84
+ "pruned_heads": {},
85
+ "remove_invalid_values": false,
86
+ "repetition_penalty": 1.0,
87
+ "return_dict": true,
88
+ "return_dict_in_generate": false,
89
+ "sep_token_id": null,
90
+ "stage_names": [
91
+ "stem",
92
+ "stage1",
93
+ "stage2",
94
+ "stage3"
95
+ ],
96
+ "suppress_tokens": null,
97
+ "task_specific_params": null,
98
+ "temperature": 1.0,
99
+ "tf_legacy_loss": false,
100
+ "tie_encoder_decoder": false,
101
+ "tie_word_embeddings": true,
102
+ "tokenizer_class": null,
103
+ "top_k": 50,
104
+ "top_p": 1.0,
105
+ "torch_dtype": null,
106
+ "torchscript": false,
107
+ "transformers_version": "4.44.2",
108
+ "typical_p": 1.0,
109
+ "use_bfloat16": false,
110
+ "width_factor": 1
111
+ },
112
+ "backbone_featmap_shape": [
113
+ 1,
114
+ 1024,
115
+ 24,
116
+ 24
117
+ ],
118
+ "backbone_kwargs": null,
119
+ "backbone_out_indices": [
120
+ 2,
121
+ 5,
122
+ 8,
123
+ 11
124
+ ],
125
+ "fusion_hidden_size": 256,
126
+ "head_in_index": -1,
127
+ "hidden_act": "gelu",
128
+ "hidden_dropout_prob": 0.0,
129
+ "hidden_size": 768,
130
+ "id2label": {
131
+ "0": "LABEL_0",
132
+ "1": "LABEL_1",
133
+ "2": "LABEL_2",
134
+ "3": "LABEL_3",
135
+ "4": "LABEL_4",
136
+ "5": "LABEL_5",
137
+ "6": "LABEL_6",
138
+ "7": "LABEL_7",
139
+ "8": "LABEL_8",
140
+ "9": "LABEL_9",
141
+ "10": "LABEL_10",
142
+ "11": "LABEL_11",
143
+ "12": "LABEL_12",
144
+ "13": "LABEL_13",
145
+ "14": "LABEL_14",
146
+ "15": "LABEL_15",
147
+ "16": "LABEL_16",
148
+ "17": "LABEL_17",
149
+ "18": "LABEL_18",
150
+ "19": "LABEL_19",
151
+ "20": "LABEL_20",
152
+ "21": "LABEL_21",
153
+ "22": "LABEL_22",
154
+ "23": "LABEL_23",
155
+ "24": "LABEL_24",
156
+ "25": "LABEL_25",
157
+ "26": "LABEL_26",
158
+ "27": "LABEL_27",
159
+ "28": "LABEL_28",
160
+ "29": "LABEL_29",
161
+ "30": "LABEL_30",
162
+ "31": "LABEL_31",
163
+ "32": "LABEL_32",
164
+ "33": "LABEL_33",
165
+ "34": "LABEL_34",
166
+ "35": "LABEL_35",
167
+ "36": "LABEL_36",
168
+ "37": "LABEL_37",
169
+ "38": "LABEL_38",
170
+ "39": "LABEL_39",
171
+ "40": "LABEL_40",
172
+ "41": "LABEL_41",
173
+ "42": "LABEL_42",
174
+ "43": "LABEL_43",
175
+ "44": "LABEL_44",
176
+ "45": "LABEL_45",
177
+ "46": "LABEL_46",
178
+ "47": "LABEL_47",
179
+ "48": "LABEL_48",
180
+ "49": "LABEL_49",
181
+ "50": "LABEL_50",
182
+ "51": "LABEL_51",
183
+ "52": "LABEL_52",
184
+ "53": "LABEL_53",
185
+ "54": "LABEL_54",
186
+ "55": "LABEL_55",
187
+ "56": "LABEL_56",
188
+ "57": "LABEL_57",
189
+ "58": "LABEL_58",
190
+ "59": "LABEL_59",
191
+ "60": "LABEL_60",
192
+ "61": "LABEL_61",
193
+ "62": "LABEL_62",
194
+ "63": "LABEL_63",
195
+ "64": "LABEL_64",
196
+ "65": "LABEL_65",
197
+ "66": "LABEL_66",
198
+ "67": "LABEL_67",
199
+ "68": "LABEL_68",
200
+ "69": "LABEL_69",
201
+ "70": "LABEL_70",
202
+ "71": "LABEL_71",
203
+ "72": "LABEL_72",
204
+ "73": "LABEL_73",
205
+ "74": "LABEL_74",
206
+ "75": "LABEL_75",
207
+ "76": "LABEL_76",
208
+ "77": "LABEL_77",
209
+ "78": "LABEL_78",
210
+ "79": "LABEL_79",
211
+ "80": "LABEL_80",
212
+ "81": "LABEL_81",
213
+ "82": "LABEL_82",
214
+ "83": "LABEL_83",
215
+ "84": "LABEL_84",
216
+ "85": "LABEL_85",
217
+ "86": "LABEL_86",
218
+ "87": "LABEL_87",
219
+ "88": "LABEL_88",
220
+ "89": "LABEL_89",
221
+ "90": "LABEL_90",
222
+ "91": "LABEL_91",
223
+ "92": "LABEL_92",
224
+ "93": "LABEL_93",
225
+ "94": "LABEL_94",
226
+ "95": "LABEL_95",
227
+ "96": "LABEL_96",
228
+ "97": "LABEL_97",
229
+ "98": "LABEL_98",
230
+ "99": "LABEL_99",
231
+ "100": "LABEL_100",
232
+ "101": "LABEL_101",
233
+ "102": "LABEL_102",
234
+ "103": "LABEL_103",
235
+ "104": "LABEL_104",
236
+ "105": "LABEL_105",
237
+ "106": "LABEL_106",
238
+ "107": "LABEL_107",
239
+ "108": "LABEL_108",
240
+ "109": "LABEL_109",
241
+ "110": "LABEL_110",
242
+ "111": "LABEL_111",
243
+ "112": "LABEL_112",
244
+ "113": "LABEL_113",
245
+ "114": "LABEL_114",
246
+ "115": "LABEL_115",
247
+ "116": "LABEL_116",
248
+ "117": "LABEL_117",
249
+ "118": "LABEL_118",
250
+ "119": "LABEL_119",
251
+ "120": "LABEL_120",
252
+ "121": "LABEL_121",
253
+ "122": "LABEL_122",
254
+ "123": "LABEL_123",
255
+ "124": "LABEL_124",
256
+ "125": "LABEL_125",
257
+ "126": "LABEL_126",
258
+ "127": "LABEL_127",
259
+ "128": "LABEL_128",
260
+ "129": "LABEL_129",
261
+ "130": "LABEL_130",
262
+ "131": "LABEL_131",
263
+ "132": "LABEL_132",
264
+ "133": "LABEL_133",
265
+ "134": "LABEL_134",
266
+ "135": "LABEL_135",
267
+ "136": "LABEL_136",
268
+ "137": "LABEL_137",
269
+ "138": "LABEL_138",
270
+ "139": "LABEL_139",
271
+ "140": "LABEL_140",
272
+ "141": "LABEL_141",
273
+ "142": "LABEL_142",
274
+ "143": "LABEL_143",
275
+ "144": "LABEL_144",
276
+ "145": "LABEL_145",
277
+ "146": "LABEL_146",
278
+ "147": "LABEL_147",
279
+ "148": "LABEL_148",
280
+ "149": "LABEL_149"
281
+ },
282
+ "image_size": 384,
283
+ "initializer_range": 0.02,
284
+ "intermediate_size": 3072,
285
+ "is_hybrid": true,
286
+ "label2id": {
287
+ "LABEL_0": 0,
288
+ "LABEL_1": 1,
289
+ "LABEL_10": 10,
290
+ "LABEL_100": 100,
291
+ "LABEL_101": 101,
292
+ "LABEL_102": 102,
293
+ "LABEL_103": 103,
294
+ "LABEL_104": 104,
295
+ "LABEL_105": 105,
296
+ "LABEL_106": 106,
297
+ "LABEL_107": 107,
298
+ "LABEL_108": 108,
299
+ "LABEL_109": 109,
300
+ "LABEL_11": 11,
301
+ "LABEL_110": 110,
302
+ "LABEL_111": 111,
303
+ "LABEL_112": 112,
304
+ "LABEL_113": 113,
305
+ "LABEL_114": 114,
306
+ "LABEL_115": 115,
307
+ "LABEL_116": 116,
308
+ "LABEL_117": 117,
309
+ "LABEL_118": 118,
310
+ "LABEL_119": 119,
311
+ "LABEL_12": 12,
312
+ "LABEL_120": 120,
313
+ "LABEL_121": 121,
314
+ "LABEL_122": 122,
315
+ "LABEL_123": 123,
316
+ "LABEL_124": 124,
317
+ "LABEL_125": 125,
318
+ "LABEL_126": 126,
319
+ "LABEL_127": 127,
320
+ "LABEL_128": 128,
321
+ "LABEL_129": 129,
322
+ "LABEL_13": 13,
323
+ "LABEL_130": 130,
324
+ "LABEL_131": 131,
325
+ "LABEL_132": 132,
326
+ "LABEL_133": 133,
327
+ "LABEL_134": 134,
328
+ "LABEL_135": 135,
329
+ "LABEL_136": 136,
330
+ "LABEL_137": 137,
331
+ "LABEL_138": 138,
332
+ "LABEL_139": 139,
333
+ "LABEL_14": 14,
334
+ "LABEL_140": 140,
335
+ "LABEL_141": 141,
336
+ "LABEL_142": 142,
337
+ "LABEL_143": 143,
338
+ "LABEL_144": 144,
339
+ "LABEL_145": 145,
340
+ "LABEL_146": 146,
341
+ "LABEL_147": 147,
342
+ "LABEL_148": 148,
343
+ "LABEL_149": 149,
344
+ "LABEL_15": 15,
345
+ "LABEL_16": 16,
346
+ "LABEL_17": 17,
347
+ "LABEL_18": 18,
348
+ "LABEL_19": 19,
349
+ "LABEL_2": 2,
350
+ "LABEL_20": 20,
351
+ "LABEL_21": 21,
352
+ "LABEL_22": 22,
353
+ "LABEL_23": 23,
354
+ "LABEL_24": 24,
355
+ "LABEL_25": 25,
356
+ "LABEL_26": 26,
357
+ "LABEL_27": 27,
358
+ "LABEL_28": 28,
359
+ "LABEL_29": 29,
360
+ "LABEL_3": 3,
361
+ "LABEL_30": 30,
362
+ "LABEL_31": 31,
363
+ "LABEL_32": 32,
364
+ "LABEL_33": 33,
365
+ "LABEL_34": 34,
366
+ "LABEL_35": 35,
367
+ "LABEL_36": 36,
368
+ "LABEL_37": 37,
369
+ "LABEL_38": 38,
370
+ "LABEL_39": 39,
371
+ "LABEL_4": 4,
372
+ "LABEL_40": 40,
373
+ "LABEL_41": 41,
374
+ "LABEL_42": 42,
375
+ "LABEL_43": 43,
376
+ "LABEL_44": 44,
377
+ "LABEL_45": 45,
378
+ "LABEL_46": 46,
379
+ "LABEL_47": 47,
380
+ "LABEL_48": 48,
381
+ "LABEL_49": 49,
382
+ "LABEL_5": 5,
383
+ "LABEL_50": 50,
384
+ "LABEL_51": 51,
385
+ "LABEL_52": 52,
386
+ "LABEL_53": 53,
387
+ "LABEL_54": 54,
388
+ "LABEL_55": 55,
389
+ "LABEL_56": 56,
390
+ "LABEL_57": 57,
391
+ "LABEL_58": 58,
392
+ "LABEL_59": 59,
393
+ "LABEL_6": 6,
394
+ "LABEL_60": 60,
395
+ "LABEL_61": 61,
396
+ "LABEL_62": 62,
397
+ "LABEL_63": 63,
398
+ "LABEL_64": 64,
399
+ "LABEL_65": 65,
400
+ "LABEL_66": 66,
401
+ "LABEL_67": 67,
402
+ "LABEL_68": 68,
403
+ "LABEL_69": 69,
404
+ "LABEL_7": 7,
405
+ "LABEL_70": 70,
406
+ "LABEL_71": 71,
407
+ "LABEL_72": 72,
408
+ "LABEL_73": 73,
409
+ "LABEL_74": 74,
410
+ "LABEL_75": 75,
411
+ "LABEL_76": 76,
412
+ "LABEL_77": 77,
413
+ "LABEL_78": 78,
414
+ "LABEL_79": 79,
415
+ "LABEL_8": 8,
416
+ "LABEL_80": 80,
417
+ "LABEL_81": 81,
418
+ "LABEL_82": 82,
419
+ "LABEL_83": 83,
420
+ "LABEL_84": 84,
421
+ "LABEL_85": 85,
422
+ "LABEL_86": 86,
423
+ "LABEL_87": 87,
424
+ "LABEL_88": 88,
425
+ "LABEL_89": 89,
426
+ "LABEL_9": 9,
427
+ "LABEL_90": 90,
428
+ "LABEL_91": 91,
429
+ "LABEL_92": 92,
430
+ "LABEL_93": 93,
431
+ "LABEL_94": 94,
432
+ "LABEL_95": 95,
433
+ "LABEL_96": 96,
434
+ "LABEL_97": 97,
435
+ "LABEL_98": 98,
436
+ "LABEL_99": 99
437
+ },
438
+ "layer_norm_eps": 1e-12,
439
+ "model_type": "dpt",
440
+ "neck_hidden_sizes": [
441
+ 256,
442
+ 512,
443
+ 768,
444
+ 768
445
+ ],
446
+ "neck_ignore_stages": [
447
+ 0,
448
+ 1
449
+ ],
450
+ "num_attention_heads": 12,
451
+ "num_channels": 3,
452
+ "num_hidden_layers": 12,
453
+ "patch_size": 16,
454
+ "qkv_bias": true,
455
+ "readout_type": "project",
456
+ "reassemble_factors": [
457
+ 1,
458
+ 1,
459
+ 1,
460
+ 0.5
461
+ ],
462
+ "semantic_classifier_dropout": 0.1,
463
+ "semantic_loss_ignore_index": 255,
464
+ "torch_dtype": "float32",
465
+ "transformers_version": null,
466
+ "use_autobackbone": false,
467
+ "use_auxiliary_head": true,
468
+ "use_batch_norm_in_fusion_residual": false,
469
+ "use_bias_in_fusion_residual": null,
470
+ "use_pretrained_backbone": false,
471
+ "use_timm_backbone": false
472
+ }
depth_estimator/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9599793d3ce64d7ebc85657360831596c1df9abc61f6820fe623fe7efb2e29c5
3
+ size 489563460
feature_extractor/preprocessor_config.json ADDED
@@ -0,0 +1,26 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_pad": false,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "ensure_multiple_of": 1,
7
+ "image_mean": [
8
+ 0.5,
9
+ 0.5,
10
+ 0.5
11
+ ],
12
+ "image_processor_type": "DPTImageProcessor",
13
+ "image_std": [
14
+ 0.5,
15
+ 0.5,
16
+ 0.5
17
+ ],
18
+ "keep_aspect_ratio": false,
19
+ "resample": 2,
20
+ "rescale_factor": 0.00392156862745098,
21
+ "size": {
22
+ "height": 384,
23
+ "width": 384
24
+ },
25
+ "size_divisor": null
26
+ }
model_index.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "StableDiffusionDepth2ImgPipeline",
3
+ "_diffusers_version": "0.30.1",
4
+ "_name_or_path": "stabilityai/stable-diffusion-2-depth",
5
+ "depth_estimator": [
6
+ "transformers",
7
+ "DPTForDepthEstimation"
8
+ ],
9
+ "feature_extractor": [
10
+ "transformers",
11
+ "DPTImageProcessor"
12
+ ],
13
+ "scheduler": [
14
+ "diffusers",
15
+ "PNDMScheduler"
16
+ ],
17
+ "text_encoder": [
18
+ "transformers",
19
+ "CLIPTextModel"
20
+ ],
21
+ "tokenizer": [
22
+ "transformers",
23
+ "CLIPTokenizer"
24
+ ],
25
+ "unet": [
26
+ "diffusers",
27
+ "UNet2DConditionModel"
28
+ ],
29
+ "vae": [
30
+ "diffusers",
31
+ "AutoencoderKL"
32
+ ]
33
+ }
scheduler/scheduler_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "PNDMScheduler",
3
+ "_diffusers_version": "0.30.1",
4
+ "beta_end": 0.012,
5
+ "beta_schedule": "scaled_linear",
6
+ "beta_start": 0.00085,
7
+ "clip_sample": false,
8
+ "num_train_timesteps": 1000,
9
+ "prediction_type": "epsilon",
10
+ "set_alpha_to_one": false,
11
+ "skip_prk_steps": true,
12
+ "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
15
+ }
text_encoder/config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "stabilityai/stable-diffusion-2-depth",
3
+ "architectures": [
4
+ "CLIPTextModel"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 0,
8
+ "dropout": 0.0,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "gelu",
11
+ "hidden_size": 1024,
12
+ "initializer_factor": 1.0,
13
+ "initializer_range": 0.02,
14
+ "intermediate_size": 4096,
15
+ "layer_norm_eps": 1e-05,
16
+ "max_position_embeddings": 77,
17
+ "model_type": "clip_text_model",
18
+ "num_attention_heads": 16,
19
+ "num_hidden_layers": 23,
20
+ "pad_token_id": 1,
21
+ "projection_dim": 512,
22
+ "torch_dtype": "float32",
23
+ "transformers_version": "4.44.2",
24
+ "vocab_size": 49415
25
+ }
text_encoder/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e024feff07dbbc87d0b6d70c549c6d618653a122497c6af15fec2fc8ede9c6e
3
+ size 1361624976
tokenizer/added_tokens.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "<back>": 49410,
3
+ "<bottom>": 49413,
4
+ "<front>": 49408,
5
+ "<left>": 49409,
6
+ "<object>": 49414,
7
+ "<overhead>": 49412,
8
+ "<right>": 49411
9
+ }
tokenizer/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer/special_tokens_map.json ADDED
@@ -0,0 +1,24 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|startoftext|>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|endoftext|>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": "!",
17
+ "unk_token": {
18
+ "content": "<|endoftext|>",
19
+ "lstrip": false,
20
+ "normalized": true,
21
+ "rstrip": false,
22
+ "single_word": false
23
+ }
24
+ }
tokenizer/tokenizer_config.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "!",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "49406": {
13
+ "content": "<|startoftext|>",
14
+ "lstrip": false,
15
+ "normalized": true,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "49407": {
21
+ "content": "<|endoftext|>",
22
+ "lstrip": false,
23
+ "normalized": true,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "49408": {
29
+ "content": "<front>",
30
+ "lstrip": false,
31
+ "normalized": true,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": false
35
+ },
36
+ "49409": {
37
+ "content": "<left>",
38
+ "lstrip": false,
39
+ "normalized": true,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": false
43
+ },
44
+ "49410": {
45
+ "content": "<back>",
46
+ "lstrip": false,
47
+ "normalized": true,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": false
51
+ },
52
+ "49411": {
53
+ "content": "<right>",
54
+ "lstrip": false,
55
+ "normalized": true,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": false
59
+ },
60
+ "49412": {
61
+ "content": "<overhead>",
62
+ "lstrip": false,
63
+ "normalized": true,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": false
67
+ },
68
+ "49413": {
69
+ "content": "<bottom>",
70
+ "lstrip": false,
71
+ "normalized": true,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": false
75
+ },
76
+ "49414": {
77
+ "content": "<object>",
78
+ "lstrip": false,
79
+ "normalized": true,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": false
83
+ }
84
+ },
85
+ "bos_token": "<|startoftext|>",
86
+ "clean_up_tokenization_spaces": true,
87
+ "do_lower_case": true,
88
+ "eos_token": "<|endoftext|>",
89
+ "errors": "replace",
90
+ "model_max_length": 77,
91
+ "pad_token": "!",
92
+ "tokenizer_class": "CLIPTokenizer",
93
+ "unk_token": "<|endoftext|>",
94
+ "use_fast": false
95
+ }
tokenizer/vocab.json ADDED
The diff for this file is too large to render. See raw diff
 
unet/config.json ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "UNet2DConditionModel",
3
+ "_diffusers_version": "0.30.1",
4
+ "_name_or_path": "stabilityai/stable-diffusion-2-depth",
5
+ "act_fn": "silu",
6
+ "addition_embed_type": null,
7
+ "addition_embed_type_num_heads": 64,
8
+ "addition_time_embed_dim": null,
9
+ "attention_head_dim": [
10
+ 5,
11
+ 10,
12
+ 20,
13
+ 20
14
+ ],
15
+ "attention_type": "default",
16
+ "block_out_channels": [
17
+ 320,
18
+ 640,
19
+ 1280,
20
+ 1280
21
+ ],
22
+ "center_input_sample": false,
23
+ "class_embed_type": null,
24
+ "class_embeddings_concat": false,
25
+ "conv_in_kernel": 3,
26
+ "conv_out_kernel": 3,
27
+ "cross_attention_dim": 1024,
28
+ "cross_attention_norm": null,
29
+ "down_block_types": [
30
+ "CrossAttnDownBlock2D",
31
+ "CrossAttnDownBlock2D",
32
+ "CrossAttnDownBlock2D",
33
+ "DownBlock2D"
34
+ ],
35
+ "downsample_padding": 1,
36
+ "dropout": 0.0,
37
+ "dual_cross_attention": false,
38
+ "encoder_hid_dim": null,
39
+ "encoder_hid_dim_type": null,
40
+ "flip_sin_to_cos": true,
41
+ "freq_shift": 0,
42
+ "in_channels": 5,
43
+ "layers_per_block": 2,
44
+ "mid_block_only_cross_attention": null,
45
+ "mid_block_scale_factor": 1,
46
+ "mid_block_type": "UNetMidBlock2DCrossAttn",
47
+ "norm_eps": 1e-05,
48
+ "norm_num_groups": 32,
49
+ "num_attention_heads": null,
50
+ "num_class_embeds": null,
51
+ "only_cross_attention": false,
52
+ "out_channels": 4,
53
+ "projection_class_embeddings_input_dim": null,
54
+ "resnet_out_scale_factor": 1.0,
55
+ "resnet_skip_time_act": false,
56
+ "resnet_time_scale_shift": "default",
57
+ "reverse_transformer_layers_per_block": null,
58
+ "sample_size": 32,
59
+ "time_cond_proj_dim": null,
60
+ "time_embedding_act_fn": null,
61
+ "time_embedding_dim": null,
62
+ "time_embedding_type": "positional",
63
+ "timestep_post_act": null,
64
+ "transformer_layers_per_block": 1,
65
+ "up_block_types": [
66
+ "UpBlock2D",
67
+ "CrossAttnUpBlock2D",
68
+ "CrossAttnUpBlock2D",
69
+ "CrossAttnUpBlock2D"
70
+ ],
71
+ "upcast_attention": false,
72
+ "use_linear_projection": true
73
+ }
unet/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:919971916c7f5371118370d2f1ab7f83618e3a3b90250b3c2eff441f9ce8e370
3
+ size 3463738032
vae/config.json ADDED
@@ -0,0 +1,38 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_class_name": "AutoencoderKL",
3
+ "_diffusers_version": "0.30.1",
4
+ "_name_or_path": "/home/techoffice/.cache/huggingface/hub/models--stabilityai--stable-diffusion-2-depth/snapshots/d49bafe6f381b0fe37ccfc4c8f6a23424b09d6ef/vae",
5
+ "act_fn": "silu",
6
+ "block_out_channels": [
7
+ 128,
8
+ 256,
9
+ 512,
10
+ 512
11
+ ],
12
+ "down_block_types": [
13
+ "DownEncoderBlock2D",
14
+ "DownEncoderBlock2D",
15
+ "DownEncoderBlock2D",
16
+ "DownEncoderBlock2D"
17
+ ],
18
+ "force_upcast": true,
19
+ "in_channels": 3,
20
+ "latent_channels": 4,
21
+ "latents_mean": null,
22
+ "latents_std": null,
23
+ "layers_per_block": 2,
24
+ "mid_block_add_attention": true,
25
+ "norm_num_groups": 32,
26
+ "out_channels": 3,
27
+ "sample_size": 256,
28
+ "scaling_factor": 0.18215,
29
+ "shift_factor": null,
30
+ "up_block_types": [
31
+ "UpDecoderBlock2D",
32
+ "UpDecoderBlock2D",
33
+ "UpDecoderBlock2D",
34
+ "UpDecoderBlock2D"
35
+ ],
36
+ "use_post_quant_conv": true,
37
+ "use_quant_conv": true
38
+ }
vae/diffusion_pytorch_model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2aa1f43011b553a4cba7f37456465cdbd48aab7b54b9348b890e8058ea7683ec
3
+ size 334643268