DipakFaceML commited on
Commit
10ca971
·
verified ·
1 Parent(s): 0173faf

Fixed model upload

Browse files
model_index.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
  "_diffusers_version": "0.37.0",
4
- "_name_or_path": "DipakFaceML/image-model",
5
  "feature_extractor": [
6
  "transformers",
7
  "CLIPImageProcessor"
@@ -10,14 +10,14 @@
10
  null,
11
  null
12
  ],
13
- "requires_safety_checker": false,
14
  "safety_checker": [
15
- null,
16
- null
17
  ],
18
  "scheduler": [
19
  "diffusers",
20
- "DPMSolverMultistepScheduler"
21
  ],
22
  "text_encoder": [
23
  "transformers",
 
1
  {
2
  "_class_name": "StableDiffusionPipeline",
3
  "_diffusers_version": "0.37.0",
4
+ "_name_or_path": "runwayml/stable-diffusion-v1-5",
5
  "feature_extractor": [
6
  "transformers",
7
  "CLIPImageProcessor"
 
10
  null,
11
  null
12
  ],
13
+ "requires_safety_checker": true,
14
  "safety_checker": [
15
+ "stable_diffusion",
16
+ "StableDiffusionSafetyChecker"
17
  ],
18
  "scheduler": [
19
  "diffusers",
20
+ "PNDMScheduler"
21
  ],
22
  "text_encoder": [
23
  "transformers",
safety_checker/config.json ADDED
@@ -0,0 +1,82 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "StableDiffusionSafetyChecker"
4
+ ],
5
+ "dtype": "float16",
6
+ "initializer_factor": 1.0,
7
+ "logit_scale_init_value": 2.6592,
8
+ "model_type": "clip",
9
+ "projection_dim": 768,
10
+ "text_config": {
11
+ "add_cross_attention": false,
12
+ "attention_dropout": 0.0,
13
+ "bos_token_id": 49406,
14
+ "cross_attention_hidden_size": null,
15
+ "decoder_start_token_id": null,
16
+ "dropout": 0.0,
17
+ "dtype": "float16",
18
+ "eos_token_id": 49407,
19
+ "finetuning_task": null,
20
+ "hidden_act": "quick_gelu",
21
+ "hidden_size": 768,
22
+ "initializer_factor": 1.0,
23
+ "initializer_range": 0.02,
24
+ "intermediate_size": 3072,
25
+ "is_decoder": false,
26
+ "layer_norm_eps": 1e-05,
27
+ "max_position_embeddings": 77,
28
+ "model_type": "clip_text_model",
29
+ "num_attention_heads": 12,
30
+ "num_hidden_layers": 12,
31
+ "pad_token_id": 1,
32
+ "prefix": null,
33
+ "projection_dim": 512,
34
+ "pruned_heads": {},
35
+ "sep_token_id": null,
36
+ "task_specific_params": null,
37
+ "tf_legacy_loss": false,
38
+ "tie_encoder_decoder": false,
39
+ "tie_word_embeddings": true,
40
+ "tokenizer_class": null,
41
+ "torchscript": false,
42
+ "use_bfloat16": false,
43
+ "vocab_size": 49408
44
+ },
45
+ "transformers_version": "5.0.0",
46
+ "vision_config": {
47
+ "add_cross_attention": false,
48
+ "attention_dropout": 0.0,
49
+ "bos_token_id": null,
50
+ "cross_attention_hidden_size": null,
51
+ "decoder_start_token_id": null,
52
+ "dropout": 0.0,
53
+ "dtype": "float16",
54
+ "eos_token_id": null,
55
+ "finetuning_task": null,
56
+ "hidden_act": "quick_gelu",
57
+ "hidden_size": 1024,
58
+ "image_size": 224,
59
+ "initializer_factor": 1.0,
60
+ "initializer_range": 0.02,
61
+ "intermediate_size": 4096,
62
+ "is_decoder": false,
63
+ "layer_norm_eps": 1e-05,
64
+ "model_type": "clip_vision_model",
65
+ "num_attention_heads": 16,
66
+ "num_channels": 3,
67
+ "num_hidden_layers": 24,
68
+ "pad_token_id": null,
69
+ "patch_size": 14,
70
+ "prefix": null,
71
+ "projection_dim": 512,
72
+ "pruned_heads": {},
73
+ "sep_token_id": null,
74
+ "task_specific_params": null,
75
+ "tf_legacy_loss": false,
76
+ "tie_encoder_decoder": false,
77
+ "tie_word_embeddings": true,
78
+ "tokenizer_class": null,
79
+ "torchscript": false,
80
+ "use_bfloat16": false
81
+ }
82
+ }
safety_checker/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:57ecdfa243b170f9b4cb3eefaf0f64552ef78fc0bf0eb1c5b9675308447184f6
3
+ size 608016280
scheduler/scheduler_config.json CHANGED
@@ -1,35 +1,15 @@
1
  {
2
- "_class_name": "DPMSolverMultistepScheduler",
3
  "_diffusers_version": "0.37.0",
4
- "algorithm_type": "dpmsolver++",
5
  "beta_end": 0.012,
6
  "beta_schedule": "scaled_linear",
7
  "beta_start": 0.00085,
8
  "clip_sample": false,
9
- "dynamic_thresholding_ratio": 0.995,
10
- "euler_at_final": false,
11
- "final_sigmas_type": "zero",
12
- "flow_shift": 1.0,
13
- "lambda_min_clipped": -Infinity,
14
- "lower_order_final": true,
15
  "num_train_timesteps": 1000,
16
  "prediction_type": "epsilon",
17
- "rescale_betas_zero_snr": false,
18
- "sample_max_value": 1.0,
19
  "set_alpha_to_one": false,
20
  "skip_prk_steps": true,
21
- "solver_order": 2,
22
- "solver_type": "midpoint",
23
  "steps_offset": 1,
24
- "thresholding": false,
25
- "time_shift_type": "exponential",
26
- "timestep_spacing": "linspace",
27
- "trained_betas": null,
28
- "use_beta_sigmas": false,
29
- "use_dynamic_shifting": false,
30
- "use_exponential_sigmas": false,
31
- "use_flow_sigmas": false,
32
- "use_karras_sigmas": false,
33
- "use_lu_lambdas": false,
34
- "variance_type": null
35
  }
 
1
  {
2
+ "_class_name": "PNDMScheduler",
3
  "_diffusers_version": "0.37.0",
 
4
  "beta_end": 0.012,
5
  "beta_schedule": "scaled_linear",
6
  "beta_start": 0.00085,
7
  "clip_sample": false,
 
 
 
 
 
 
8
  "num_train_timesteps": 1000,
9
  "prediction_type": "epsilon",
 
 
10
  "set_alpha_to_one": false,
11
  "skip_prk_steps": true,
 
 
12
  "steps_offset": 1,
13
+ "timestep_spacing": "leading",
14
+ "trained_betas": null
 
 
 
 
 
 
 
 
 
15
  }
unet/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
  "_diffusers_version": "0.37.0",
4
- "_name_or_path": "/root/.cache/huggingface/hub/models--DipakFaceML--image-model/snapshots/08d0b33ced2c11be155df628996881d380335a42/unet",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
7
  "addition_embed_type_num_heads": 64,
 
1
  {
2
  "_class_name": "UNet2DConditionModel",
3
  "_diffusers_version": "0.37.0",
4
+ "_name_or_path": "/root/.cache/huggingface/hub/models--runwayml--stable-diffusion-v1-5/snapshots/451f4fe16113bff5a5d2269ed5ad43b0592e9a14/unet",
5
  "act_fn": "silu",
6
  "addition_embed_type": null,
7
  "addition_embed_type_num_heads": 64,
unet/diffusion_pytorch_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ff3fd745b5135f4d6fa72a1b1a20e0079fa511014f17b611b4f9e8695dcaf1c6
3
  size 1719125304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c83908253f9a64d08c25fc90874c9c8aef9a329ce1ca5fb909d73b0c83d1ea21
3
  size 1719125304
vae/config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "_class_name": "AutoencoderKL",
3
  "_diffusers_version": "0.37.0",
4
- "_name_or_path": "/root/.cache/huggingface/hub/models--DipakFaceML--image-model/snapshots/08d0b33ced2c11be155df628996881d380335a42/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,
 
1
  {
2
  "_class_name": "AutoencoderKL",
3
  "_diffusers_version": "0.37.0",
4
+ "_name_or_path": "/root/.cache/huggingface/hub/models--runwayml--stable-diffusion-v1-5/snapshots/451f4fe16113bff5a5d2269ed5ad43b0592e9a14/vae",
5
  "act_fn": "silu",
6
  "block_out_channels": [
7
  128,