ArrayCats commited on
Commit
80c58bb
·
verified ·
1 Parent(s): 09126c5

Upload 222 files

Browse files
Files changed (44) hide show
  1. .gitattributes +23 -0
  2. A-表番角色/IS〈Infinite Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.civitai.info +893 -0
  3. A-表番角色/IS〈Infinite Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.preview.png +3 -0
  4. A-表番角色/IS〈Infinite Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.safetensors +3 -0
  5. A-表番角色/IS〈Infinite Stratos〉/Char-IS-Cecilia-V1-SD1.5.civitai.info +910 -0
  6. A-表番角色/IS〈Infinite Stratos〉/Char-IS-Cecilia-V1-SD1.5.preview.png +3 -0
  7. A-表番角色/IS〈Infinite Stratos〉/Char-IS-Cecilia-V1-SD1.5.safetensors +3 -0
  8. A-表番角色/噬血狂袭/15_Aiba_Asagi-30.civitai.info +232 -0
  9. A-表番角色/噬血狂袭/15_Aiba_Asagi-30.preview.png +3 -0
  10. A-表番角色/噬血狂袭/15_Aiba_Asagi-30.safetensors +3 -0
  11. A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.civitai.info +185 -0
  12. A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.preview.png +3 -0
  13. A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.safetensors +3 -0
  14. A-表番角色/噬血狂袭/15_Astarte-30.civitai.info +232 -0
  15. A-表番角色/噬血狂袭/15_Astarte-30.preview.png +3 -0
  16. A-表番角色/噬血狂袭/15_Astarte-30.safetensors +3 -0
  17. A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.civitai.info +232 -0
  18. A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.preview.png +3 -0
  19. A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.safetensors +3 -0
  20. A-表番角色/噬血狂袭/15_Kanase_Kanon-30.civitai.info +232 -0
  21. A-表番角色/噬血狂袭/15_Kanase_Kanon-30.preview.png +3 -0
  22. A-表番角色/噬血狂袭/15_Kanase_Kanon-30.safetensors +3 -0
  23. A-表番角色/噬血狂袭/KirasakaSayakaStrike_v10.preview.png +0 -0
  24. A-表番角色/噬血狂袭/asagi_aiba.preview.png +0 -0
  25. A-表番角色/噬血狂袭/kanaseKanonStrikeTheBlood_v10.preview.png +0 -0
  26. A-表番角色/噬血狂袭/la_folia.preview.png +0 -0
  27. A-表番角色/噬血狂袭/natsuki_minamiya.preview.png +0 -0
  28. A-表番角色/噬血狂袭/natsuki_minamiya_no_outfit.preview.png +0 -0
  29. A-表番角色/噬血狂袭/yukina_himeragi.preview.png +0 -0
  30. A-表番角色/噬血狂袭/yukina_himeragi_no_outfit.preview.png +0 -0
  31. A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.civitai.info +291 -0
  32. A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.preview.png +3 -0
  33. A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.safetensors +3 -0
  34. A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.civitai.info +293 -0
  35. A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.preview.png +3 -0
  36. A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.safetensors +3 -0
  37. A-表番角色/精灵幻想记/aishia.preview.png +0 -0
  38. A-表番角色/精灵幻想记/dryas.preview.png +0 -0
  39. A-表番角色/精灵幻想记/orphia.preview.png +0 -0
  40. A-表番角色/精灵幻想记/orphia_no_outfit.preview.png +0 -0
  41. A-表番角色/精灵幻想记/精霊幻想記莎拉(サラ).preview.png +0 -0
  42. A-表番角色/高达系列/AmiaLee-10.civitai.info +179 -0
  43. A-表番角色/高达系列/AmiaLee-10.preview.png +3 -0
  44. A-表番角色/高达系列/AmiaLee-10.safetensors +3 -0
.gitattributes CHANGED
@@ -1394,3 +1394,26 @@ A-表番角色/孤独摇滚/Otsuki_Yoyoko-v5.preview.png filter=lfs diff=lfs mer
1394
  A-表番角色/孤独摇滚/pa-san-10.preview.png filter=lfs diff=lfs merge=lfs -text
1395
  A-表番角色/为了养老,我要在异世界存8万枚金币/aderet_300_d32a32_v2_nobg.preview.png filter=lfs diff=lfs merge=lfs -text
1396
  A-表番角色/英雄王,为了穷尽武道而转生~而后成为世界最强见习骑士♀~/leone_eiyuu_v1b(resize-sv10).preview.png filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1394
  A-表番角色/孤独摇滚/pa-san-10.preview.png filter=lfs diff=lfs merge=lfs -text
1395
  A-表番角色/为了养老,我要在异世界存8万枚金币/aderet_300_d32a32_v2_nobg.preview.png filter=lfs diff=lfs merge=lfs -text
1396
  A-表番角色/英雄王,为了穷尽武道而转生~而后成为世界最强见习骑士♀~/leone_eiyuu_v1b(resize-sv10).preview.png filter=lfs diff=lfs merge=lfs -text
1397
+ A-表番角色/高达系列/AmiaLee-10.preview.png filter=lfs diff=lfs merge=lfs -text
1398
+ A-表番角色/精灵幻想记/精霊幻想記莎拉(サラ).preview.png filter=lfs diff=lfs merge=lfs -text
1399
+ A-表番角色/精灵幻想记/aishia.preview.png filter=lfs diff=lfs merge=lfs -text
1400
+ A-表番角色/精灵幻想记/dryas.preview.png filter=lfs diff=lfs merge=lfs -text
1401
+ A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.preview.png filter=lfs diff=lfs merge=lfs -text
1402
+ A-表番角色/精灵幻想记/orphia_no_outfit.preview.png filter=lfs diff=lfs merge=lfs -text
1403
+ A-表番角色/精灵幻想记/orphia.preview.png filter=lfs diff=lfs merge=lfs -text
1404
+ A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.preview.png filter=lfs diff=lfs merge=lfs -text
1405
+ A-表番角色/噬血狂袭/15_Aiba_Asagi-30.preview.png filter=lfs diff=lfs merge=lfs -text
1406
+ A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.preview.png filter=lfs diff=lfs merge=lfs -text
1407
+ A-表番角色/噬血狂袭/15_Astarte-30.preview.png filter=lfs diff=lfs merge=lfs -text
1408
+ A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.preview.png filter=lfs diff=lfs merge=lfs -text
1409
+ A-表番角色/噬血狂袭/15_Kanase_Kanon-30.preview.png filter=lfs diff=lfs merge=lfs -text
1410
+ A-表番角色/噬血狂袭/asagi_aiba.preview.png filter=lfs diff=lfs merge=lfs -text
1411
+ A-表番角色/噬血狂袭/kanaseKanonStrikeTheBlood_v10.preview.png filter=lfs diff=lfs merge=lfs -text
1412
+ A-表番角色/噬血狂袭/KirasakaSayakaStrike_v10.preview.png filter=lfs diff=lfs merge=lfs -text
1413
+ A-表番角色/噬血狂袭/la_folia.preview.png filter=lfs diff=lfs merge=lfs -text
1414
+ A-表番角色/噬血狂袭/natsuki_minamiya_no_outfit.preview.png filter=lfs diff=lfs merge=lfs -text
1415
+ A-表番角色/噬血狂袭/natsuki_minamiya.preview.png filter=lfs diff=lfs merge=lfs -text
1416
+ A-表番角色/噬血狂袭/yukina_himeragi_no_outfit.preview.png filter=lfs diff=lfs merge=lfs -text
1417
+ A-表番角色/噬血狂袭/yukina_himeragi.preview.png filter=lfs diff=lfs merge=lfs -text
1418
+ A-表番角色/IS〈Infinite[[:space:]]Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.preview.png filter=lfs diff=lfs merge=lfs -text
1419
+ A-表番角色/IS〈Infinite[[:space:]]Stratos〉/Char-IS-Cecilia-V1-SD1.5.preview.png filter=lfs diff=lfs merge=lfs -text
A-表番角色/IS〈Infinite Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.civitai.info ADDED
@@ -0,0 +1,893 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2207993,
3
+ "modelId": 1950899,
4
+ "name": "v1.0-SD1.5",
5
+ "nsfwLevel": 5,
6
+ "createdAt": "2025-09-12T13:44:27.643Z",
7
+ "updatedAt": "2025-09-25T07:25:58.843Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-09-14T11:00:00.000Z",
10
+ "trainedWords": [
11
+ "charlotte dunois",
12
+ "low ponytail, hair ribbon, solo, blue neck ribbon, white coat dress, white collared shirt, sneakers",
13
+ "braided ponytail, ankh necklace, yellow bikini, striped bikini skirt"
14
+ ],
15
+ "trainingStatus": null,
16
+ "trainingDetails": null,
17
+ "baseModel": "SD 1.5",
18
+ "baseModelType": "Standard",
19
+ "earlyAccessEndsAt": null,
20
+ "earlyAccessConfig": null,
21
+ "description": "Initial release of SD1.5 version, trained with 1024x1024 resolution, and with all training data from original anime",
22
+ "uploadType": "Created",
23
+ "usageControl": "Download",
24
+ "air": "urn:air:sd1:lora:civitai:1950899@2207993",
25
+ "stats": {
26
+ "downloadCount": 92,
27
+ "ratingCount": 0,
28
+ "rating": 0,
29
+ "thumbsUpCount": 14
30
+ },
31
+ "model": {
32
+ "name": "Charlotte Dunois",
33
+ "type": "LORA",
34
+ "nsfw": false,
35
+ "poi": false,
36
+ "description": "AboutThis is the Lora for the character Charlotte Dunois (\u30b7\u30e3\u30eb\u30ed\u30c3\u30c8\u30fb\u30c7\u30e5\u30ce\u30a2) of Infinite Stratos.\n\nIL, Pony version was trained with 1536x1536, and SDXL and SD1.5 version was trained with 1024x1024 resolution\n\nTrained with original TV Anime, offering the most authentic visual of the story.\n\n\n\nUsageTrigger word: charlotte dunois\n\nFor default JK costume: low ponytail, hair ribbon, solo, blue neck ribbon, white coat dress, white collared shirt, sneakers\n\nFor The iconic bikini: braided ponytail, ankh necklace, yellow bikini, striped bikini skirt",
37
+ "tags": [
38
+ "charlotte dunois",
39
+ "character",
40
+ "infinite stratos"
41
+ ],
42
+ "allowNoCredit": false,
43
+ "allowCommercialUse": [
44
+ "Image",
45
+ "RentCivit",
46
+ "Rent"
47
+ ],
48
+ "allowDerivatives": false,
49
+ "allowDifferentLicense": false
50
+ },
51
+ "files": [
52
+ {
53
+ "id": 2100946,
54
+ "sizeKB": 46991.82421875,
55
+ "name": "Char-Charlotte_Dunois-V1-SD1.5.safetensors",
56
+ "type": "Model",
57
+ "pickleScanResult": "Success",
58
+ "pickleScanMessage": "No Pickle imports",
59
+ "virusScanResult": "Success",
60
+ "virusScanMessage": null,
61
+ "scannedAt": "2025-09-12T13:46:07.369Z",
62
+ "metadata": {
63
+ "format": "SafeTensor",
64
+ "size": null,
65
+ "fp": null
66
+ },
67
+ "hashes": {
68
+ "AutoV1": "17BDE8F7",
69
+ "AutoV2": "DBAF133F31",
70
+ "SHA256": "DBAF133F317A9FE48E4A7DC52344DC1C35E725A721984784D4089AC48AEDBCB6",
71
+ "CRC32": "53B9A139",
72
+ "BLAKE3": "05187F531C21E0AC2C589E0267D877786F8862E2F9B4159A498C31BD5F1AFBE0",
73
+ "AutoV3": "D1FDE17EA452"
74
+ },
75
+ "primary": true,
76
+ "downloadUrl": "https://civitai.com/api/download/models/2207993"
77
+ }
78
+ ],
79
+ "images": [
80
+ {
81
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/21e11b17-cbb8-4d67-8bb9-36c58bad6ff3/original=true/99886589.jpeg",
82
+ "nsfwLevel": 1,
83
+ "width": 1280,
84
+ "height": 1920,
85
+ "hash": "UFLD#{~95XIt_NxYyGae_3WGo%$*Bq-oSds,",
86
+ "type": "image",
87
+ "metadata": {
88
+ "hash": "UFLD#{~95XIt_NxYyGae_3WGo%$*Bq-oSds,",
89
+ "size": 2591939,
90
+ "width": 1280,
91
+ "height": 1920
92
+ },
93
+ "minor": false,
94
+ "poi": false,
95
+ "meta": {
96
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
97
+ "Size": "640x960",
98
+ "seed": 3410381545,
99
+ "Model": "1Gen-AM-Any6Cet3Bra1",
100
+ "steps": 40,
101
+ "hashes": {
102
+ "vae": "f921fb3f29",
103
+ "model": "a3d3cc400d",
104
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
105
+ },
106
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, low ponytail, hair ribbon, solo, blue neck ribbon, white coat dress, white collared shirt, looking at viewer, smug, open mouth, naughty face, hand on own face, standing, cowboy shot, outdoors, street, cherry blossoms, petals, depth of field",
107
+ "Version": "v1.10.1",
108
+ "sampler": "Euler a",
109
+ "cfgScale": 7,
110
+ "Pad conds": "True",
111
+ "resources": [
112
+ {
113
+ "hash": "0xff8eb3af5f",
114
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
115
+ "type": "lora"
116
+ },
117
+ {
118
+ "hash": "a3d3cc400d",
119
+ "name": "1Gen-AM-Any6Cet3Bra1",
120
+ "type": "model"
121
+ }
122
+ ],
123
+ "Model hash": "a3d3cc400d",
124
+ "Hires steps": "30",
125
+ "Hires upscale": "2",
126
+ "Schedule type": "Karras",
127
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
128
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
129
+ "ADetailer model": "face_yolov8n.pt",
130
+ "ADetailer steps": "90",
131
+ "ADetailer sampler": "Euler a",
132
+ "ADetailer version": "25.3.0",
133
+ "Denoising strength": "0.35",
134
+ "ADetailer mask blur": "4",
135
+ "ADetailer model 3rd": "hand_yolov8n.pt",
136
+ "ADetailer scheduler": "SGM Uniform",
137
+ "ADetailer confidence": "0.3",
138
+ "ADetailer prompt 3rd": {
139
+ "GoodHands-beta2": "1>\\nmasterpiece"
140
+ },
141
+ "ADetailer dilate erode": "4",
142
+ "ADetailer mask blur 3rd": "4",
143
+ "ADetailer confidence 3rd": "0.82",
144
+ "ADetailer inpaint padding": "32",
145
+ "ADetailer dilate erode 3rd": "4",
146
+ "ADetailer denoising strength": "0.4",
147
+ "ADetailer use separate steps": "True",
148
+ "ADetailer inpaint only masked": "True",
149
+ "ADetailer inpaint padding 3rd": "32",
150
+ "ADetailer use separate sampler": "True",
151
+ "ADetailer denoising strength 3rd": "0.4",
152
+ "ADetailer inpaint only masked 3rd": "True"
153
+ },
154
+ "availability": "Public",
155
+ "hasMeta": true,
156
+ "hasPositivePrompt": true,
157
+ "onSite": false,
158
+ "remixOfId": null
159
+ },
160
+ {
161
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/34511ff8-9398-47a7-8265-47b2a9cffbf4/original=true/99886597.jpeg",
162
+ "nsfwLevel": 1,
163
+ "width": 1280,
164
+ "height": 1920,
165
+ "hash": "UEHnjGIp9b0g_N-oNe?a01IUx]Ne00xYE258",
166
+ "type": "image",
167
+ "metadata": {
168
+ "hash": "UEHnjGIp9b0g_N-oNe?a01IUx]Ne00xYE258",
169
+ "size": 2108857,
170
+ "width": 1280,
171
+ "height": 1920
172
+ },
173
+ "minor": false,
174
+ "poi": false,
175
+ "meta": {
176
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
177
+ "Size": "640x960",
178
+ "seed": 3879572556,
179
+ "Model": "1Gen-AM-Any6Cet3Bra1",
180
+ "steps": 40,
181
+ "hashes": {
182
+ "vae": "f921fb3f29",
183
+ "model": "a3d3cc400d",
184
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
185
+ },
186
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, low ponytail, hair ribbon, solo, blue neck ribbon, blue neck ribbonwhite coat dress, white collared shirt, sneakers, standing, contrapposto, full body, looking at viewer, smile, open mouth, sunset, classroom, school desk, window, backlighting, depth of field",
187
+ "Version": "v1.10.1",
188
+ "sampler": "Euler a",
189
+ "cfgScale": 7,
190
+ "Pad conds": "True",
191
+ "resources": [
192
+ {
193
+ "hash": "0xff8eb3af5f",
194
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
195
+ "type": "lora"
196
+ },
197
+ {
198
+ "hash": "a3d3cc400d",
199
+ "name": "1Gen-AM-Any6Cet3Bra1",
200
+ "type": "model"
201
+ }
202
+ ],
203
+ "Model hash": "a3d3cc400d",
204
+ "Hires steps": "30",
205
+ "Hires upscale": "2",
206
+ "Schedule type": "Karras",
207
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
208
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
209
+ "ADetailer model": "face_yolov8n.pt",
210
+ "ADetailer steps": "90",
211
+ "ADetailer sampler": "Euler a",
212
+ "ADetailer version": "25.3.0",
213
+ "Denoising strength": "0.35",
214
+ "ADetailer mask blur": "4",
215
+ "ADetailer model 3rd": "hand_yolov8n.pt",
216
+ "ADetailer scheduler": "SGM Uniform",
217
+ "ADetailer confidence": "0.3",
218
+ "ADetailer prompt 3rd": {
219
+ "GoodHands-beta2": "1>\\nmasterpiece"
220
+ },
221
+ "ADetailer dilate erode": "4",
222
+ "ADetailer mask blur 3rd": "4",
223
+ "ADetailer confidence 3rd": "0.82",
224
+ "ADetailer inpaint padding": "32",
225
+ "ADetailer dilate erode 3rd": "4",
226
+ "ADetailer denoising strength": "0.4",
227
+ "ADetailer use separate steps": "True",
228
+ "ADetailer inpaint only masked": "True",
229
+ "ADetailer inpaint padding 3rd": "32",
230
+ "ADetailer use separate sampler": "True",
231
+ "ADetailer denoising strength 3rd": "0.4",
232
+ "ADetailer inpaint only masked 3rd": "True"
233
+ },
234
+ "availability": "Public",
235
+ "hasMeta": true,
236
+ "hasPositivePrompt": true,
237
+ "onSite": false,
238
+ "remixOfId": null
239
+ },
240
+ {
241
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/4369a9bf-b5b2-42bf-a098-2f94bde2748e/original=true/99886595.jpeg",
242
+ "nsfwLevel": 1,
243
+ "width": 1280,
244
+ "height": 1920,
245
+ "hash": "UBMG;nz.0n_2PX=^O]Di?a=xO[R557M_01-.",
246
+ "type": "image",
247
+ "metadata": {
248
+ "hash": "UBMG;nz.0n_2PX=^O]Di?a=xO[R557M_01-.",
249
+ "size": 2886561,
250
+ "width": 1280,
251
+ "height": 1920
252
+ },
253
+ "minor": false,
254
+ "poi": false,
255
+ "meta": {
256
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
257
+ "Size": "640x960",
258
+ "seed": 5082747,
259
+ "Model": "REV-I",
260
+ "steps": 40,
261
+ "hashes": {
262
+ "vae": "f921fb3f29",
263
+ "model": "3f6c5087ad",
264
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
265
+ },
266
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, low ponytail, hair ribbon, solo, blue neck ribbon, white coat dress, white collared shirt, waving, looking at viewer, smile, open mouth, standing, cowboy shot, outdoors, street, cherry blossoms, petals, depth of field",
267
+ "Version": "v1.10.1",
268
+ "sampler": "Euler a",
269
+ "cfgScale": 7,
270
+ "Pad conds": "True",
271
+ "resources": [
272
+ {
273
+ "hash": "0xff8eb3af5f",
274
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
275
+ "type": "lora"
276
+ },
277
+ {
278
+ "hash": "3f6c5087ad",
279
+ "name": "REV-I",
280
+ "type": "model"
281
+ }
282
+ ],
283
+ "Model hash": "3f6c5087ad",
284
+ "Hires steps": "30",
285
+ "Hires upscale": "2",
286
+ "Schedule type": "Karras",
287
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
288
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
289
+ "ADetailer model": "face_yolov8n.pt",
290
+ "ADetailer steps": "90",
291
+ "ADetailer sampler": "Euler a",
292
+ "ADetailer version": "25.3.0",
293
+ "Denoising strength": "0.35",
294
+ "ADetailer mask blur": "4",
295
+ "ADetailer model 3rd": "hand_yolov8n.pt",
296
+ "ADetailer scheduler": "SGM Uniform",
297
+ "ADetailer confidence": "0.3",
298
+ "ADetailer prompt 3rd": {
299
+ "GoodHands-beta2": "1>\\nmasterpiece"
300
+ },
301
+ "ADetailer dilate erode": "4",
302
+ "ADetailer mask blur 3rd": "4",
303
+ "ADetailer confidence 3rd": "0.82",
304
+ "ADetailer inpaint padding": "32",
305
+ "ADetailer dilate erode 3rd": "4",
306
+ "ADetailer denoising strength": "0.4",
307
+ "ADetailer use separate steps": "True",
308
+ "ADetailer inpaint only masked": "True",
309
+ "ADetailer inpaint padding 3rd": "32",
310
+ "ADetailer use separate sampler": "True",
311
+ "ADetailer denoising strength 3rd": "0.4",
312
+ "ADetailer inpaint only masked 3rd": "True"
313
+ },
314
+ "availability": "Public",
315
+ "hasMeta": true,
316
+ "hasPositivePrompt": true,
317
+ "onSite": false,
318
+ "remixOfId": null
319
+ },
320
+ {
321
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/4e01e69b-9ed7-4647-9cc4-f0a4a80bac61/original=true/99886604.jpeg",
322
+ "nsfwLevel": 1,
323
+ "width": 1280,
324
+ "height": 1920,
325
+ "hash": "UHKwqY?vEQESyZEg%OS$o~9tohWAF2v~ivNH",
326
+ "type": "image",
327
+ "metadata": {
328
+ "hash": "UHKwqY?vEQESyZEg%OS$o~9tohWAF2v~ivNH",
329
+ "size": 2594302,
330
+ "width": 1280,
331
+ "height": 1920
332
+ },
333
+ "minor": false,
334
+ "poi": false,
335
+ "meta": {
336
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
337
+ "Size": "640x960",
338
+ "seed": 3762658343,
339
+ "Model": "1Gen-AM-Any6Cet3Bra1",
340
+ "steps": 40,
341
+ "hashes": {
342
+ "vae": "f921fb3f29",
343
+ "model": "a3d3cc400d",
344
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
345
+ },
346
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, low ponytail, hair ribbon, solo, blue neck ribbon, white coat dress, white collared shirt, hand on hip, looking at viewer, light smile, standing, cowboy shot, outdoors, street, cherry blossoms, petals, depth of field",
347
+ "Version": "v1.10.1",
348
+ "sampler": "Euler a",
349
+ "cfgScale": 7,
350
+ "Pad conds": "True",
351
+ "resources": [
352
+ {
353
+ "hash": "0xff8eb3af5f",
354
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
355
+ "type": "lora"
356
+ },
357
+ {
358
+ "hash": "a3d3cc400d",
359
+ "name": "1Gen-AM-Any6Cet3Bra1",
360
+ "type": "model"
361
+ }
362
+ ],
363
+ "Model hash": "a3d3cc400d",
364
+ "Hires steps": "30",
365
+ "Hires upscale": "2",
366
+ "Schedule type": "Karras",
367
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
368
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
369
+ "ADetailer model": "face_yolov8n.pt",
370
+ "ADetailer steps": "90",
371
+ "ADetailer sampler": "Euler a",
372
+ "ADetailer version": "25.3.0",
373
+ "Denoising strength": "0.35",
374
+ "ADetailer mask blur": "4",
375
+ "ADetailer model 3rd": "hand_yolov8n.pt",
376
+ "ADetailer scheduler": "SGM Uniform",
377
+ "ADetailer confidence": "0.3",
378
+ "ADetailer prompt 3rd": {
379
+ "GoodHands-beta2": "1>\\nmasterpiece"
380
+ },
381
+ "ADetailer dilate erode": "4",
382
+ "ADetailer mask blur 3rd": "4",
383
+ "ADetailer confidence 3rd": "0.82",
384
+ "ADetailer inpaint padding": "32",
385
+ "ADetailer dilate erode 3rd": "4",
386
+ "ADetailer denoising strength": "0.4",
387
+ "ADetailer use separate steps": "True",
388
+ "ADetailer inpaint only masked": "True",
389
+ "ADetailer inpaint padding 3rd": "32",
390
+ "ADetailer use separate sampler": "True",
391
+ "ADetailer denoising strength 3rd": "0.4",
392
+ "ADetailer inpaint only masked 3rd": "True"
393
+ },
394
+ "availability": "Public",
395
+ "hasMeta": true,
396
+ "hasPositivePrompt": true,
397
+ "onSite": false,
398
+ "remixOfId": null
399
+ },
400
+ {
401
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/779fcfad-560a-43fc-b1bc-0a6f6f6e7f50/original=true/99886793.jpeg",
402
+ "nsfwLevel": 1,
403
+ "width": 1280,
404
+ "height": 1920,
405
+ "hash": "UIH2Zj~q00VrvJRQ0,M{?Y%2-q.7x]jvNIWB",
406
+ "type": "image",
407
+ "metadata": {
408
+ "hash": "UIH2Zj~q00VrvJRQ0,M{?Y%2-q.7x]jvNIWB",
409
+ "size": 2396479,
410
+ "width": 1280,
411
+ "height": 1920
412
+ },
413
+ "minor": false,
414
+ "poi": false,
415
+ "meta": {
416
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
417
+ "Size": "640x960",
418
+ "seed": 2684108525,
419
+ "Model": "REV-I",
420
+ "steps": 40,
421
+ "hashes": {
422
+ "vae": "f921fb3f29",
423
+ "model": "3f6c5087ad",
424
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
425
+ },
426
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, braided ponytail, solo, kimono, hair flower, obi, sash, tabi, standing, cowboy shot, hands on own chest, light smile, east asian architecture, outdoors, blurry background, masterpiece",
427
+ "Version": "v1.10.1",
428
+ "sampler": "Euler a",
429
+ "cfgScale": 7,
430
+ "Pad conds": "True",
431
+ "resources": [
432
+ {
433
+ "hash": "0xff8eb3af5f",
434
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
435
+ "type": "lora"
436
+ },
437
+ {
438
+ "hash": "3f6c5087ad",
439
+ "name": "REV-I",
440
+ "type": "model"
441
+ }
442
+ ],
443
+ "Model hash": "3f6c5087ad",
444
+ "Hires steps": "30",
445
+ "Hires upscale": "2",
446
+ "Schedule type": "Karras",
447
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
448
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
449
+ "ADetailer model": "face_yolov8n.pt",
450
+ "ADetailer steps": "90",
451
+ "ADetailer sampler": "Euler a",
452
+ "ADetailer version": "25.3.0",
453
+ "Denoising strength": "0.35",
454
+ "ADetailer mask blur": "4",
455
+ "ADetailer model 3rd": "hand_yolov8n.pt",
456
+ "ADetailer scheduler": "SGM Uniform",
457
+ "ADetailer confidence": "0.3",
458
+ "ADetailer prompt 3rd": {
459
+ "GoodHands-beta2": "1>\\nmasterpiece"
460
+ },
461
+ "ADetailer dilate erode": "4",
462
+ "ADetailer mask blur 3rd": "4",
463
+ "ADetailer confidence 3rd": "0.82",
464
+ "ADetailer inpaint padding": "32",
465
+ "ADetailer dilate erode 3rd": "4",
466
+ "ADetailer denoising strength": "0.4",
467
+ "ADetailer use separate steps": "True",
468
+ "ADetailer inpaint only masked": "True",
469
+ "ADetailer inpaint padding 3rd": "32",
470
+ "ADetailer use separate sampler": "True",
471
+ "ADetailer denoising strength 3rd": "0.4",
472
+ "ADetailer inpaint only masked 3rd": "True"
473
+ },
474
+ "availability": "Public",
475
+ "hasMeta": true,
476
+ "hasPositivePrompt": true,
477
+ "onSite": false,
478
+ "remixOfId": null
479
+ },
480
+ {
481
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/c4357542-687f-43b2-92f9-9948a9c9c402/original=true/99886791.jpeg",
482
+ "nsfwLevel": 1,
483
+ "width": 1280,
484
+ "height": 1920,
485
+ "hash": "UGHdjR~A00Dj;0xZtlIpWAjFI@R.~BxZIVnP",
486
+ "type": "image",
487
+ "metadata": {
488
+ "hash": "UGHdjR~A00Dj;0xZtlIpWAjFI@R.~BxZIVnP",
489
+ "size": 2394502,
490
+ "width": 1280,
491
+ "height": 1920
492
+ },
493
+ "minor": false,
494
+ "poi": false,
495
+ "meta": {
496
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
497
+ "Size": "640x960",
498
+ "seed": 2243360564,
499
+ "Model": "1Gen-AM-Any6Cet3Bra1",
500
+ "steps": 40,
501
+ "hashes": {
502
+ "vae": "f921fb3f29",
503
+ "model": "a3d3cc400d",
504
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
505
+ },
506
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, solo, hair down, long hair, maid, maid headdress, maid apron, embarrassed, white pantyhose, open mouth, blush, parted lips, looking at viewer, cowboy shot, bar \\(place\\), indoors, depth of field",
507
+ "Version": "v1.10.1",
508
+ "sampler": "Euler a",
509
+ "cfgScale": 7,
510
+ "Pad conds": "True",
511
+ "resources": [
512
+ {
513
+ "hash": "0xff8eb3af5f",
514
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
515
+ "type": "lora"
516
+ },
517
+ {
518
+ "hash": "a3d3cc400d",
519
+ "name": "1Gen-AM-Any6Cet3Bra1",
520
+ "type": "model"
521
+ }
522
+ ],
523
+ "Model hash": "a3d3cc400d",
524
+ "Hires steps": "30",
525
+ "Hires upscale": "2",
526
+ "Schedule type": "Karras",
527
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
528
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
529
+ "ADetailer model": "face_yolov8n.pt",
530
+ "ADetailer steps": "90",
531
+ "ADetailer sampler": "Euler a",
532
+ "ADetailer version": "25.3.0",
533
+ "Denoising strength": "0.35",
534
+ "ADetailer mask blur": "4",
535
+ "ADetailer model 3rd": "hand_yolov8n.pt",
536
+ "ADetailer scheduler": "SGM Uniform",
537
+ "ADetailer confidence": "0.3",
538
+ "ADetailer prompt 3rd": {
539
+ "GoodHands-beta2": "1>\\nmasterpiece"
540
+ },
541
+ "ADetailer dilate erode": "4",
542
+ "ADetailer mask blur 3rd": "4",
543
+ "ADetailer confidence 3rd": "0.82",
544
+ "ADetailer inpaint padding": "32",
545
+ "ADetailer dilate erode 3rd": "4",
546
+ "ADetailer denoising strength": "0.4",
547
+ "ADetailer use separate steps": "True",
548
+ "ADetailer inpaint only masked": "True",
549
+ "ADetailer inpaint padding 3rd": "32",
550
+ "ADetailer use separate sampler": "True",
551
+ "ADetailer denoising strength 3rd": "0.4",
552
+ "ADetailer inpaint only masked 3rd": "True"
553
+ },
554
+ "availability": "Public",
555
+ "hasMeta": true,
556
+ "hasPositivePrompt": true,
557
+ "onSite": false,
558
+ "remixOfId": null
559
+ },
560
+ {
561
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/658ef36c-cb4c-450f-a9dc-57f0094dd29f/original=true/99886792.jpeg",
562
+ "nsfwLevel": 4,
563
+ "width": 1280,
564
+ "height": 1920,
565
+ "hash": "UGIgu^Ip00={~2tRXARPRh%MtRNGItWB~Uof",
566
+ "type": "image",
567
+ "metadata": {
568
+ "hash": "UGIgu^Ip00={~2tRXARPRh%MtRNGItWB~Uof",
569
+ "size": 2324734,
570
+ "width": 1280,
571
+ "height": 1920
572
+ },
573
+ "minor": false,
574
+ "poi": false,
575
+ "meta": {
576
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
577
+ "Size": "640x960",
578
+ "seed": 1621785444,
579
+ "Model": "1Gen-AM-Any6Cet3Bra1",
580
+ "steps": 40,
581
+ "hashes": {
582
+ "vae": "f921fb3f29",
583
+ "model": "a3d3cc400d",
584
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
585
+ },
586
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, solo, hair down, long hair, solo, playboy bunny, fishnets, fake animal ears, rabbit ears, bar \\(place\\), indoors, depth of field, cowboy shot, smirk",
587
+ "Version": "v1.10.1",
588
+ "sampler": "Euler a",
589
+ "cfgScale": 7,
590
+ "Pad conds": "True",
591
+ "resources": [
592
+ {
593
+ "hash": "0xff8eb3af5f",
594
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
595
+ "type": "lora"
596
+ },
597
+ {
598
+ "hash": "a3d3cc400d",
599
+ "name": "1Gen-AM-Any6Cet3Bra1",
600
+ "type": "model"
601
+ }
602
+ ],
603
+ "Model hash": "a3d3cc400d",
604
+ "Hires steps": "30",
605
+ "Hires upscale": "2",
606
+ "Schedule type": "Karras",
607
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
608
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
609
+ "ADetailer model": "face_yolov8n.pt",
610
+ "ADetailer steps": "90",
611
+ "ADetailer sampler": "Euler a",
612
+ "ADetailer version": "25.3.0",
613
+ "Denoising strength": "0.35",
614
+ "ADetailer mask blur": "4",
615
+ "ADetailer model 3rd": "hand_yolov8n.pt",
616
+ "ADetailer scheduler": "SGM Uniform",
617
+ "ADetailer confidence": "0.3",
618
+ "ADetailer prompt 3rd": {
619
+ "GoodHands-beta2": "1>\\nmasterpiece"
620
+ },
621
+ "ADetailer dilate erode": "4",
622
+ "ADetailer mask blur 3rd": "4",
623
+ "ADetailer confidence 3rd": "0.82",
624
+ "ADetailer inpaint padding": "32",
625
+ "ADetailer dilate erode 3rd": "4",
626
+ "ADetailer denoising strength": "0.4",
627
+ "ADetailer use separate steps": "True",
628
+ "ADetailer inpaint only masked": "True",
629
+ "ADetailer inpaint padding 3rd": "32",
630
+ "ADetailer use separate sampler": "True",
631
+ "ADetailer denoising strength 3rd": "0.4",
632
+ "ADetailer inpaint only masked 3rd": "True"
633
+ },
634
+ "availability": "Public",
635
+ "hasMeta": true,
636
+ "hasPositivePrompt": true,
637
+ "onSite": false,
638
+ "remixOfId": null
639
+ },
640
+ {
641
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/678d44ef-efd3-424f-9f48-c6122156b123/original=true/99886790.jpeg",
642
+ "nsfwLevel": 4,
643
+ "width": 1280,
644
+ "height": 1920,
645
+ "hash": "UMH2$XNG00xt.RozNuM|J~Rjtlof_NS4RPs:",
646
+ "type": "image",
647
+ "metadata": {
648
+ "hash": "UMH2$XNG00xt.RozNuM|J~Rjtlof_NS4RPs:",
649
+ "size": 2359303,
650
+ "width": 1280,
651
+ "height": 1920
652
+ },
653
+ "minor": false,
654
+ "poi": false,
655
+ "meta": {
656
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
657
+ "Size": "640x960",
658
+ "seed": 2381835988,
659
+ "Model": "1Gen-AM-Any6Cet3Bra1",
660
+ "steps": 40,
661
+ "hashes": {
662
+ "vae": "f921fb3f29",
663
+ "model": "a3d3cc400d",
664
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
665
+ },
666
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, solo, charlotte dunois, solo, braided ponytail, (sundress, white dress, sleeveless dress:1.2), bare legs, stream, wading, (skirt hold, curtsey: 1.2), looking at viewer, light smile, forest, depth of field",
667
+ "Version": "v1.10.1",
668
+ "sampler": "Euler a",
669
+ "cfgScale": 7,
670
+ "Pad conds": "True",
671
+ "resources": [
672
+ {
673
+ "hash": "0xff8eb3af5f",
674
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
675
+ "type": "lora"
676
+ },
677
+ {
678
+ "hash": "a3d3cc400d",
679
+ "name": "1Gen-AM-Any6Cet3Bra1",
680
+ "type": "model"
681
+ }
682
+ ],
683
+ "Model hash": "a3d3cc400d",
684
+ "Hires steps": "30",
685
+ "Hires upscale": "2",
686
+ "Schedule type": "Karras",
687
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
688
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
689
+ "ADetailer model": "face_yolov8n.pt",
690
+ "ADetailer steps": "90",
691
+ "ADetailer sampler": "Euler a",
692
+ "ADetailer version": "25.3.0",
693
+ "Denoising strength": "0.35",
694
+ "ADetailer mask blur": "4",
695
+ "ADetailer model 3rd": "hand_yolov8n.pt",
696
+ "ADetailer scheduler": "SGM Uniform",
697
+ "ADetailer confidence": "0.3",
698
+ "ADetailer prompt 3rd": {
699
+ "GoodHands-beta2": "1>\\nmasterpiece"
700
+ },
701
+ "ADetailer dilate erode": "4",
702
+ "ADetailer mask blur 3rd": "4",
703
+ "ADetailer confidence 3rd": "0.82",
704
+ "ADetailer inpaint padding": "32",
705
+ "ADetailer dilate erode 3rd": "4",
706
+ "ADetailer denoising strength": "0.4",
707
+ "ADetailer use separate steps": "True",
708
+ "ADetailer inpaint only masked": "True",
709
+ "ADetailer inpaint padding 3rd": "32",
710
+ "ADetailer use separate sampler": "True",
711
+ "ADetailer denoising strength 3rd": "0.4",
712
+ "ADetailer inpaint only masked 3rd": "True"
713
+ },
714
+ "availability": "Public",
715
+ "hasMeta": true,
716
+ "hasPositivePrompt": true,
717
+ "onSite": false,
718
+ "remixOfId": null
719
+ },
720
+ {
721
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/8d6e198e-6dfe-4a5e-b378-b120edc99ec6/original=true/99886849.jpeg",
722
+ "nsfwLevel": 4,
723
+ "width": 1280,
724
+ "height": 1920,
725
+ "hash": "UPJ[YS-=3G0.NK-ojJS%Mwo~-5bbN#ohD*wJ",
726
+ "type": "image",
727
+ "metadata": {
728
+ "hash": "UPJ[YS-=3G0.NK-ojJS%Mwo~-5bbN#ohD*wJ",
729
+ "size": 2672873,
730
+ "width": 1280,
731
+ "height": 1920
732
+ },
733
+ "minor": false,
734
+ "poi": false,
735
+ "meta": {
736
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
737
+ "Size": "640x960",
738
+ "seed": 1789769057,
739
+ "Model": "1Gen-AM-Any6Cet3Bra1",
740
+ "steps": 40,
741
+ "hashes": {
742
+ "vae": "f921fb3f29",
743
+ "model": "a3d3cc400d",
744
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
745
+ },
746
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, solo, braided ponytail, ankh necklace, yellow bikini, striped bikini skirt, cowboy shot, looking at viewer, beach, ocean, sunlight, blue sky, depth of field, laughing",
747
+ "Version": "v1.10.1",
748
+ "sampler": "Euler a",
749
+ "cfgScale": 7,
750
+ "Pad conds": "True",
751
+ "resources": [
752
+ {
753
+ "hash": "0xff8eb3af5f",
754
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
755
+ "type": "lora"
756
+ },
757
+ {
758
+ "hash": "a3d3cc400d",
759
+ "name": "1Gen-AM-Any6Cet3Bra1",
760
+ "type": "model"
761
+ }
762
+ ],
763
+ "Model hash": "a3d3cc400d",
764
+ "Hires steps": "30",
765
+ "Hires upscale": "2",
766
+ "Schedule type": "Karras",
767
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
768
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
769
+ "ADetailer model": "face_yolov8n.pt",
770
+ "ADetailer steps": "90",
771
+ "ADetailer sampler": "Euler a",
772
+ "ADetailer version": "25.3.0",
773
+ "Denoising strength": "0.35",
774
+ "ADetailer mask blur": "4",
775
+ "ADetailer model 3rd": "hand_yolov8n.pt",
776
+ "ADetailer scheduler": "SGM Uniform",
777
+ "ADetailer confidence": "0.3",
778
+ "ADetailer prompt 3rd": {
779
+ "GoodHands-beta2": "1>\\nmasterpiece"
780
+ },
781
+ "ADetailer dilate erode": "4",
782
+ "ADetailer mask blur 3rd": "4",
783
+ "ADetailer confidence 3rd": "0.82",
784
+ "ADetailer inpaint padding": "32",
785
+ "ADetailer dilate erode 3rd": "4",
786
+ "ADetailer denoising strength": "0.4",
787
+ "ADetailer use separate steps": "True",
788
+ "ADetailer inpaint only masked": "True",
789
+ "ADetailer inpaint padding 3rd": "32",
790
+ "ADetailer use separate sampler": "True",
791
+ "ADetailer denoising strength 3rd": "0.4",
792
+ "ADetailer inpaint only masked 3rd": "True"
793
+ },
794
+ "availability": "Public",
795
+ "hasMeta": true,
796
+ "hasPositivePrompt": true,
797
+ "onSite": false,
798
+ "remixOfId": null
799
+ },
800
+ {
801
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d3f2e8ea-8470-44d4-b385-655ac9f71bc7/original=true/99886852.jpeg",
802
+ "nsfwLevel": 4,
803
+ "width": 1280,
804
+ "height": 1920,
805
+ "hash": "UQKBgzxt2|F5Fg-QJDIurtR+#SxaNLt7oIad",
806
+ "type": "image",
807
+ "metadata": {
808
+ "hash": "UQKBgzxt2|F5Fg-QJDIurtR+#SxaNLt7oIad",
809
+ "size": 2517320,
810
+ "width": 1280,
811
+ "height": 1920
812
+ },
813
+ "minor": false,
814
+ "poi": false,
815
+ "meta": {
816
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
817
+ "Size": "640x960",
818
+ "seed": 1831365302,
819
+ "Model": "1Gen-AM-Any6Cet3Bra1",
820
+ "steps": 40,
821
+ "hashes": {
822
+ "vae": "f921fb3f29",
823
+ "model": "a3d3cc400d",
824
+ "lora:Char-Charlotte_Dunois-V1-SD1.5": "0xff8eb3af5f"
825
+ },
826
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-Charlotte_Dunois-V1-SD1.5:0.9:hr=0.6>, 1girl, charlotte dunois, solo, braided ponytail, ankh necklace, yellow bikini, striped bikini skirt, looking at viewer, reclining, beach, ocean, sunlight, blue sky, depth of field, laughing",
827
+ "Version": "v1.10.1",
828
+ "sampler": "Euler a",
829
+ "cfgScale": 7,
830
+ "Pad conds": "True",
831
+ "resources": [
832
+ {
833
+ "hash": "0xff8eb3af5f",
834
+ "name": "Char-Charlotte_Dunois-V1-SD1.5",
835
+ "type": "lora"
836
+ },
837
+ {
838
+ "hash": "a3d3cc400d",
839
+ "name": "1Gen-AM-Any6Cet3Bra1",
840
+ "type": "model"
841
+ }
842
+ ],
843
+ "Model hash": "a3d3cc400d",
844
+ "Hires steps": "30",
845
+ "Hires upscale": "2",
846
+ "Schedule type": "Karras",
847
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
848
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
849
+ "ADetailer model": "face_yolov8n.pt",
850
+ "ADetailer steps": "90",
851
+ "ADetailer sampler": "Euler a",
852
+ "ADetailer version": "25.3.0",
853
+ "Denoising strength": "0.35",
854
+ "ADetailer mask blur": "4",
855
+ "ADetailer model 3rd": "hand_yolov8n.pt",
856
+ "ADetailer scheduler": "SGM Uniform",
857
+ "ADetailer confidence": "0.3",
858
+ "ADetailer prompt 3rd": {
859
+ "GoodHands-beta2": "1>\\nmasterpiece"
860
+ },
861
+ "ADetailer dilate erode": "4",
862
+ "ADetailer mask blur 3rd": "4",
863
+ "ADetailer confidence 3rd": "0.82",
864
+ "ADetailer inpaint padding": "32",
865
+ "ADetailer dilate erode 3rd": "4",
866
+ "ADetailer denoising strength": "0.4",
867
+ "ADetailer use separate steps": "True",
868
+ "ADetailer inpaint only masked": "True",
869
+ "ADetailer inpaint padding 3rd": "32",
870
+ "ADetailer use separate sampler": "True",
871
+ "ADetailer denoising strength 3rd": "0.4",
872
+ "ADetailer inpaint only masked 3rd": "True"
873
+ },
874
+ "availability": "Public",
875
+ "hasMeta": true,
876
+ "hasPositivePrompt": true,
877
+ "onSite": false,
878
+ "remixOfId": null
879
+ }
880
+ ],
881
+ "downloadUrl": "https://civitai.com/api/download/models/2207993",
882
+ "creator": {
883
+ "username": "okingjo",
884
+ "image": "https://avatars.githubusercontent.com/u/25336857?v=4"
885
+ },
886
+ "extensions": {
887
+ "sd_civitai_helper": {
888
+ "version": "1.8.13",
889
+ "last_update": 1760877749,
890
+ "skeleton_file": false
891
+ }
892
+ }
893
+ }
A-表番角色/IS〈Infinite Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.preview.png ADDED

Git LFS Details

  • SHA256: ffead54410442930037ccb26842f349fc99de09cd6daac78728aaaf5f8578884
  • Pointer size: 132 Bytes
  • Size of remote file: 2.59 MB
A-表番角色/IS〈Infinite Stratos〉/Char-Charlotte_Dunois-V1-SD1.5.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dbaf133f317a9fe48e4a7dc52344dc1c35e725a721984784d4089ac48aedbcb6
3
+ size 48119628
A-表番角色/IS〈Infinite Stratos〉/Char-IS-Cecilia-V1-SD1.5.civitai.info ADDED
@@ -0,0 +1,910 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2121554,
3
+ "modelId": 1874382,
4
+ "name": "v1.0-SD1.5",
5
+ "nsfwLevel": 31,
6
+ "createdAt": "2025-08-17T06:27:56.604Z",
7
+ "updatedAt": "2025-09-02T04:18:13.992Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-08-17T11:00:00.000Z",
10
+ "trainedWords": [
11
+ "cecilia alcott",
12
+ "blue hairband, blue neck ribbon, infinite stratos academy school uniform, black pantyhose, loafers"
13
+ ],
14
+ "trainingStatus": null,
15
+ "trainingDetails": null,
16
+ "baseModel": "SD 1.5",
17
+ "baseModelType": "Standard",
18
+ "earlyAccessEndsAt": null,
19
+ "earlyAccessConfig": null,
20
+ "description": "Initial release of SD1.5 version, trained with 1024x1024 resolution, and with all training data from original anime",
21
+ "uploadType": "Created",
22
+ "usageControl": "Download",
23
+ "air": "urn:air:sd1:lora:civitai:1874382@2121554",
24
+ "stats": {
25
+ "downloadCount": 109,
26
+ "ratingCount": 0,
27
+ "rating": 0,
28
+ "thumbsUpCount": 23
29
+ },
30
+ "model": {
31
+ "name": "Cecilia Alcott (IL,Pony,XL,1.5)",
32
+ "type": "LORA",
33
+ "nsfw": false,
34
+ "poi": false,
35
+ "description": "AboutThis is the Lora for the character Cecilia Alcott (\u30bb\u30b7\u30ea\u30a2\u30fb\u30aa\u30eb\u30b3\u30c3\u30c8) of Infinite Stratos.\n\nIL, Pony version was trained with 1536x1536, and SDXL and SD1.5 version was trained with 1024x1024 resolution\n\nTrained with original TV Anime, offering the most authentic visual of the story.\n\n\n\nUsageTrigger word: cecilia alcott\n\nFor default JK costume: blue hairband, blue neck ribbon, infinite stratos academy school uniform, black pantyhose, loafers",
36
+ "tags": [
37
+ "cecilia alcott",
38
+ "character",
39
+ "infinite stratos"
40
+ ],
41
+ "allowNoCredit": false,
42
+ "allowCommercialUse": [
43
+ "Image",
44
+ "RentCivit",
45
+ "Rent"
46
+ ],
47
+ "allowDerivatives": true,
48
+ "allowDifferentLicense": false
49
+ },
50
+ "files": [
51
+ {
52
+ "id": 2015727,
53
+ "sizeKB": 46980.62109375,
54
+ "name": "Char-IS-Cecilia-V1-SD1.5.safetensors",
55
+ "type": "Model",
56
+ "pickleScanResult": "Success",
57
+ "pickleScanMessage": "No Pickle imports",
58
+ "virusScanResult": "Success",
59
+ "virusScanMessage": null,
60
+ "scannedAt": "2025-08-17T06:30:38.782Z",
61
+ "metadata": {
62
+ "format": "SafeTensor",
63
+ "size": null,
64
+ "fp": null
65
+ },
66
+ "hashes": {
67
+ "AutoV1": "9E3AA28D",
68
+ "AutoV2": "EBB7885F02",
69
+ "SHA256": "EBB7885F028FC959281D613847C5D51CB8BABA18CA2289039FC9C6B501E8AC2E",
70
+ "CRC32": "2A6CFE8E",
71
+ "BLAKE3": "BAF69EA17C005D7768C819C853422A68B83CD777455AB4F59FDB40CFCF8B932E",
72
+ "AutoV3": "7FC3E385EEEC"
73
+ },
74
+ "primary": true,
75
+ "downloadUrl": "https://civitai.com/api/download/models/2121554"
76
+ }
77
+ ],
78
+ "images": [
79
+ {
80
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0367adf5-d3a8-4769-ab68-109eb5346881/original=true/94763150.jpeg",
81
+ "nsfwLevel": 1,
82
+ "width": 1280,
83
+ "height": 1920,
84
+ "hash": "UQKn6+wbIVx]_4-U%3W?o$sm-pM|Ny%MIToL",
85
+ "type": "image",
86
+ "metadata": {
87
+ "hash": "UQKn6+wbIVx]_4-U%3W?o$sm-pM|Ny%MIToL",
88
+ "size": 2690708,
89
+ "width": 1280,
90
+ "height": 1920
91
+ },
92
+ "minor": false,
93
+ "poi": false,
94
+ "meta": {
95
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
96
+ "Size": "640x960",
97
+ "seed": 1738371536,
98
+ "Model": "spiritforeseermix_spiritforeseerO",
99
+ "steps": 40,
100
+ "hashes": {
101
+ "vae": "f921fb3f29",
102
+ "model": "1e0c69b67c",
103
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
104
+ },
105
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, blue hairband, solo, blue neck ribbon, infinite stratos academy school uniform, black pantyhose, double v, looking at viewer, smile, open mouth, standing, cowboy shot, outdoors, street, cherry blossoms, petals, depth of field",
106
+ "Version": "v1.10.1",
107
+ "sampler": "Euler a",
108
+ "cfgScale": 7,
109
+ "resources": [
110
+ {
111
+ "hash": "0x4eb7db2367",
112
+ "name": "Char-IS-Cecilia-V1-SD1.5",
113
+ "type": "lora"
114
+ },
115
+ {
116
+ "hash": "1e0c69b67c",
117
+ "name": "spiritforeseermix_spiritforeseerO",
118
+ "type": "model"
119
+ }
120
+ ],
121
+ "Model hash": "1e0c69b67c",
122
+ "Hires steps": "30",
123
+ "Hires upscale": "2",
124
+ "Schedule type": "Karras",
125
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
126
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
127
+ "ADetailer model": "face_yolov8n.pt",
128
+ "ADetailer steps": "90",
129
+ "ADetailer sampler": "Euler a",
130
+ "ADetailer version": "25.3.0",
131
+ "Denoising strength": "0.35",
132
+ "ADetailer mask blur": "4",
133
+ "ADetailer model 3rd": "hand_yolov8n.pt",
134
+ "ADetailer scheduler": "SGM Uniform",
135
+ "ADetailer confidence": "0.3",
136
+ "ADetailer prompt 3rd": {
137
+ "GoodHands-beta2": "1>\\nmasterpiece"
138
+ },
139
+ "ADetailer dilate erode": "4",
140
+ "ADetailer mask blur 3rd": "4",
141
+ "ADetailer confidence 3rd": "0.82",
142
+ "ADetailer inpaint padding": "32",
143
+ "ADetailer dilate erode 3rd": "4",
144
+ "ADetailer denoising strength": "0.4",
145
+ "ADetailer use separate steps": "True",
146
+ "ADetailer inpaint only masked": "True",
147
+ "ADetailer inpaint padding 3rd": "32",
148
+ "ADetailer use separate sampler": "True",
149
+ "ADetailer denoising strength 3rd": "0.4",
150
+ "ADetailer inpaint only masked 3rd": "True"
151
+ },
152
+ "availability": "Public",
153
+ "hasMeta": true,
154
+ "hasPositivePrompt": true,
155
+ "onSite": false,
156
+ "remixOfId": null
157
+ },
158
+ {
159
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/375e71ed-5e6f-4b92-948d-deb8ae316a07/original=true/94763147.jpeg",
160
+ "nsfwLevel": 1,
161
+ "width": 1280,
162
+ "height": 1920,
163
+ "hash": "UEHxNhZ~01I[^,4.E1tRiixG~Woz%5jXs:of",
164
+ "type": "image",
165
+ "metadata": {
166
+ "hash": "UEHxNhZ~01I[^,4.E1tRiixG~Woz%5jXs:of",
167
+ "size": 2578275,
168
+ "width": 1280,
169
+ "height": 1920
170
+ },
171
+ "minor": false,
172
+ "poi": false,
173
+ "meta": {
174
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
175
+ "Size": "640x960",
176
+ "seed": 2113766323,
177
+ "Model": "1Gen-AM-Any6Cet3Bra1",
178
+ "steps": 40,
179
+ "hashes": {
180
+ "vae": "f921fb3f29",
181
+ "model": "a3d3cc400d",
182
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
183
+ },
184
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, blue hairband, solo, blue neck ribbon, infinite stratos academy school uniform, black pantyhose, loafers, waving, looking at viewer, smile, open mouth, standing, full body, outdoors, street, cherry blossoms, petals, depth of field",
185
+ "Version": "v1.10.1",
186
+ "sampler": "Euler a",
187
+ "cfgScale": 7,
188
+ "resources": [
189
+ {
190
+ "hash": "0x4eb7db2367",
191
+ "name": "Char-IS-Cecilia-V1-SD1.5",
192
+ "type": "lora"
193
+ },
194
+ {
195
+ "hash": "a3d3cc400d",
196
+ "name": "1Gen-AM-Any6Cet3Bra1",
197
+ "type": "model"
198
+ }
199
+ ],
200
+ "Model hash": "a3d3cc400d",
201
+ "Hires steps": "30",
202
+ "Hires upscale": "2",
203
+ "Schedule type": "Karras",
204
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
205
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
206
+ "ADetailer model": "face_yolov8n.pt",
207
+ "ADetailer steps": "90",
208
+ "ADetailer sampler": "Euler a",
209
+ "ADetailer version": "25.3.0",
210
+ "Denoising strength": "0.35",
211
+ "ADetailer mask blur": "4",
212
+ "ADetailer model 3rd": "hand_yolov8n.pt",
213
+ "ADetailer scheduler": "SGM Uniform",
214
+ "ADetailer confidence": "0.3",
215
+ "ADetailer prompt 3rd": {
216
+ "GoodHands-beta2": "1>\\nmasterpiece"
217
+ },
218
+ "ADetailer dilate erode": "4",
219
+ "ADetailer mask blur 3rd": "4",
220
+ "ADetailer confidence 3rd": "0.82",
221
+ "ADetailer inpaint padding": "32",
222
+ "ADetailer dilate erode 3rd": "4",
223
+ "ADetailer denoising strength": "0.4",
224
+ "ADetailer use separate steps": "True",
225
+ "ADetailer inpaint only masked": "True",
226
+ "ADetailer inpaint padding 3rd": "32",
227
+ "ADetailer use separate sampler": "True",
228
+ "ADetailer denoising strength 3rd": "0.4",
229
+ "ADetailer inpaint only masked 3rd": "True"
230
+ },
231
+ "availability": "Public",
232
+ "hasMeta": true,
233
+ "hasPositivePrompt": true,
234
+ "onSite": false,
235
+ "remixOfId": null
236
+ },
237
+ {
238
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/a4542608-54b1-4089-bef3-cbe7c43af9c8/original=true/94763151.jpeg",
239
+ "nsfwLevel": 1,
240
+ "width": 1280,
241
+ "height": 1920,
242
+ "hash": "UFHdmcoe00SO#%57R-soE3I@?HxZ~Ba}IUt6",
243
+ "type": "image",
244
+ "metadata": {
245
+ "hash": "UFHdmcoe00SO#%57R-soE3I@?HxZ~Ba}IUt6",
246
+ "size": 2232502,
247
+ "width": 1280,
248
+ "height": 1920
249
+ },
250
+ "minor": false,
251
+ "poi": false,
252
+ "meta": {
253
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
254
+ "Size": "640x960",
255
+ "seed": 2985954694,
256
+ "Model": "1Gen-AM-Any6Cet3Bra1",
257
+ "steps": 40,
258
+ "hashes": {
259
+ "vae": "f921fb3f29",
260
+ "model": "a3d3cc400d",
261
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
262
+ },
263
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, blue hairband, solo, blue neck ribbon, infinite stratos academy school uniform, black pantyhose, looking at viewer, seductive smile, ecstasy, gasping, finger to cheek, standing, cowboy shot, bedroom, indoors, blurry background",
264
+ "Version": "v1.10.1",
265
+ "sampler": "Euler a",
266
+ "cfgScale": 7,
267
+ "resources": [
268
+ {
269
+ "hash": "0x4eb7db2367",
270
+ "name": "Char-IS-Cecilia-V1-SD1.5",
271
+ "type": "lora"
272
+ },
273
+ {
274
+ "hash": "a3d3cc400d",
275
+ "name": "1Gen-AM-Any6Cet3Bra1",
276
+ "type": "model"
277
+ }
278
+ ],
279
+ "Model hash": "a3d3cc400d",
280
+ "Hires steps": "30",
281
+ "Hires upscale": "2",
282
+ "Schedule type": "Karras",
283
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
284
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
285
+ "ADetailer model": "face_yolov8n.pt",
286
+ "ADetailer steps": "90",
287
+ "ADetailer sampler": "Euler a",
288
+ "ADetailer version": "25.3.0",
289
+ "Denoising strength": "0.35",
290
+ "ADetailer mask blur": "4",
291
+ "ADetailer model 3rd": "hand_yolov8n.pt",
292
+ "ADetailer scheduler": "SGM Uniform",
293
+ "ADetailer confidence": "0.3",
294
+ "ADetailer prompt 3rd": {
295
+ "GoodHands-beta2": "1>\\nmasterpiece"
296
+ },
297
+ "ADetailer dilate erode": "4",
298
+ "ADetailer mask blur 3rd": "4",
299
+ "ADetailer confidence 3rd": "0.82",
300
+ "ADetailer inpaint padding": "32",
301
+ "ADetailer dilate erode 3rd": "4",
302
+ "ADetailer denoising strength": "0.4",
303
+ "ADetailer use separate steps": "True",
304
+ "ADetailer inpaint only masked": "True",
305
+ "ADetailer inpaint padding 3rd": "32",
306
+ "ADetailer use separate sampler": "True",
307
+ "ADetailer denoising strength 3rd": "0.4",
308
+ "ADetailer inpaint only masked 3rd": "True"
309
+ },
310
+ "availability": "Public",
311
+ "hasMeta": true,
312
+ "hasPositivePrompt": true,
313
+ "onSite": false,
314
+ "remixOfId": null
315
+ },
316
+ {
317
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/7274cb73-a9ee-4d9b-82bc-2e19c62bf71f/original=true/94763153.jpeg",
318
+ "nsfwLevel": 1,
319
+ "width": 1280,
320
+ "height": 1920,
321
+ "hash": "USINt0~BxIIV%3s:S5n%9ZX9kB%LMfN_sloz",
322
+ "type": "image",
323
+ "metadata": {
324
+ "hash": "USINt0~BxIIV%3s:S5n%9ZX9kB%LMfN_sloz",
325
+ "size": 2481504,
326
+ "width": 1280,
327
+ "height": 1920
328
+ },
329
+ "minor": false,
330
+ "poi": false,
331
+ "meta": {
332
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
333
+ "Size": "640x960",
334
+ "seed": 3110970867,
335
+ "Model": "1Gen-AM-Any6Cet3Bra1",
336
+ "steps": 40,
337
+ "hashes": {
338
+ "vae": "f921fb3f29",
339
+ "model": "a3d3cc400d",
340
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
341
+ },
342
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, blue hairband, solo, blue neck ribbon, infinite stratos academy school uniform, black pantyhose, loafers, double v, looking at viewer, smile, open mouth, standing, full body, outdoors, street, cherry blossoms, petals, depth of field",
343
+ "Version": "v1.10.1",
344
+ "sampler": "Euler a",
345
+ "cfgScale": 7,
346
+ "resources": [
347
+ {
348
+ "hash": "0x4eb7db2367",
349
+ "name": "Char-IS-Cecilia-V1-SD1.5",
350
+ "type": "lora"
351
+ },
352
+ {
353
+ "hash": "a3d3cc400d",
354
+ "name": "1Gen-AM-Any6Cet3Bra1",
355
+ "type": "model"
356
+ }
357
+ ],
358
+ "Model hash": "a3d3cc400d",
359
+ "Hires steps": "30",
360
+ "Hires upscale": "2",
361
+ "Schedule type": "Karras",
362
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
363
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
364
+ "ADetailer model": "face_yolov8n.pt",
365
+ "ADetailer steps": "90",
366
+ "ADetailer sampler": "Euler a",
367
+ "ADetailer version": "25.3.0",
368
+ "Denoising strength": "0.35",
369
+ "ADetailer mask blur": "4",
370
+ "ADetailer model 3rd": "hand_yolov8n.pt",
371
+ "ADetailer scheduler": "SGM Uniform",
372
+ "ADetailer confidence": "0.3",
373
+ "ADetailer prompt 3rd": {
374
+ "GoodHands-beta2": "1>\\nmasterpiece"
375
+ },
376
+ "ADetailer dilate erode": "4",
377
+ "ADetailer mask blur 3rd": "4",
378
+ "ADetailer confidence 3rd": "0.82",
379
+ "ADetailer inpaint padding": "32",
380
+ "ADetailer dilate erode 3rd": "4",
381
+ "ADetailer denoising strength": "0.4",
382
+ "ADetailer use separate steps": "True",
383
+ "ADetailer inpaint only masked": "True",
384
+ "ADetailer inpaint padding 3rd": "32",
385
+ "ADetailer use separate sampler": "True",
386
+ "ADetailer denoising strength 3rd": "0.4",
387
+ "ADetailer inpaint only masked 3rd": "True"
388
+ },
389
+ "availability": "Public",
390
+ "hasMeta": true,
391
+ "hasPositivePrompt": true,
392
+ "onSite": false,
393
+ "remixOfId": null
394
+ },
395
+ {
396
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e46c8cd5-f44b-4873-8f7f-40744e0d738e/original=true/94763282.jpeg",
397
+ "nsfwLevel": 1,
398
+ "width": 1280,
399
+ "height": 1920,
400
+ "hash": "U6IE;EMy009DEGro9|ET?q8_~X-?yZ9ZMc4o",
401
+ "type": "image",
402
+ "metadata": {
403
+ "hash": "U6IE;EMy009DEGro9|ET?q8_~X-?yZ9ZMc4o",
404
+ "size": 2449494,
405
+ "width": 1280,
406
+ "height": 1920
407
+ },
408
+ "minor": false,
409
+ "poi": false,
410
+ "meta": {
411
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
412
+ "Size": "640x960",
413
+ "seed": 3014254535,
414
+ "Model": "spiritforeseermix_spiritforeseerO",
415
+ "steps": 40,
416
+ "hashes": {
417
+ "vae": "f921fb3f29",
418
+ "model": "1e0c69b67c",
419
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
420
+ },
421
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, blue hairband, solo, kimono, hair flower, obi, sash, tabi, zouri, standing, full body, hands on own chest, light smile, east asian architecture, outdoors, blurry background",
422
+ "Version": "v1.10.1",
423
+ "sampler": "Euler a",
424
+ "cfgScale": 7,
425
+ "resources": [
426
+ {
427
+ "hash": "0x4eb7db2367",
428
+ "name": "Char-IS-Cecilia-V1-SD1.5",
429
+ "type": "lora"
430
+ },
431
+ {
432
+ "hash": "1e0c69b67c",
433
+ "name": "spiritforeseermix_spiritforeseerO",
434
+ "type": "model"
435
+ }
436
+ ],
437
+ "Model hash": "1e0c69b67c",
438
+ "Hires steps": "30",
439
+ "Hires upscale": "2",
440
+ "Schedule type": "Karras",
441
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
442
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
443
+ "ADetailer model": "face_yolov8n.pt",
444
+ "ADetailer steps": "90",
445
+ "ADetailer sampler": "Euler a",
446
+ "ADetailer version": "25.3.0",
447
+ "Denoising strength": "0.35",
448
+ "ADetailer mask blur": "4",
449
+ "ADetailer model 3rd": "hand_yolov8n.pt",
450
+ "ADetailer scheduler": "SGM Uniform",
451
+ "ADetailer confidence": "0.3",
452
+ "ADetailer prompt 3rd": {
453
+ "GoodHands-beta2": "1>\\nmasterpiece"
454
+ },
455
+ "ADetailer dilate erode": "4",
456
+ "ADetailer mask blur 3rd": "4",
457
+ "ADetailer confidence 3rd": "0.82",
458
+ "ADetailer inpaint padding": "32",
459
+ "ADetailer dilate erode 3rd": "4",
460
+ "ADetailer denoising strength": "0.4",
461
+ "ADetailer use separate steps": "True",
462
+ "ADetailer inpaint only masked": "True",
463
+ "ADetailer inpaint padding 3rd": "32",
464
+ "ADetailer use separate sampler": "True",
465
+ "ADetailer denoising strength 3rd": "0.4",
466
+ "ADetailer inpaint only masked 3rd": "True"
467
+ },
468
+ "availability": "Public",
469
+ "hasMeta": true,
470
+ "hasPositivePrompt": true,
471
+ "onSite": false,
472
+ "remixOfId": null
473
+ },
474
+ {
475
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d4b20f24-e2b4-4c56-b255-935e87809122/original=true/94763285.jpeg",
476
+ "nsfwLevel": 2,
477
+ "width": 1280,
478
+ "height": 1920,
479
+ "hash": "UCIg.d~A00nn00E2JCxZ0zX9-;-okCxZngju",
480
+ "type": "image",
481
+ "metadata": {
482
+ "hash": "UCIg.d~A00nn00E2JCxZ0zX9-;-okCxZngju",
483
+ "size": 2316460,
484
+ "width": 1280,
485
+ "height": 1920
486
+ },
487
+ "minor": false,
488
+ "poi": false,
489
+ "meta": {
490
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
491
+ "Size": "640x960",
492
+ "seed": 1641057567,
493
+ "Model": "spiritforeseermix_spiritforeseerO",
494
+ "steps": 40,
495
+ "hashes": {
496
+ "vae": "f921fb3f29",
497
+ "model": "1e0c69b67c",
498
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
499
+ },
500
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, blue hairband, solo, white elbow gloves, necklace, black evening gown, light smile, ball room, indoors, depth of field, cowboy shot, sitting, chair, flower, holding cup, wine glass",
501
+ "Version": "v1.10.1",
502
+ "sampler": "Euler a",
503
+ "cfgScale": 7,
504
+ "resources": [
505
+ {
506
+ "hash": "0x4eb7db2367",
507
+ "name": "Char-IS-Cecilia-V1-SD1.5",
508
+ "type": "lora"
509
+ },
510
+ {
511
+ "hash": "1e0c69b67c",
512
+ "name": "spiritforeseermix_spiritforeseerO",
513
+ "type": "model"
514
+ }
515
+ ],
516
+ "Model hash": "1e0c69b67c",
517
+ "Hires steps": "30",
518
+ "Hires upscale": "2",
519
+ "Schedule type": "Karras",
520
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
521
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
522
+ "ADetailer model": "face_yolov8n.pt",
523
+ "ADetailer steps": "90",
524
+ "ADetailer sampler": "Euler a",
525
+ "ADetailer version": "25.3.0",
526
+ "Denoising strength": "0.35",
527
+ "ADetailer mask blur": "4",
528
+ "ADetailer model 3rd": "hand_yolov8n.pt",
529
+ "ADetailer scheduler": "SGM Uniform",
530
+ "ADetailer confidence": "0.3",
531
+ "ADetailer prompt 3rd": {
532
+ "GoodHands-beta2": "1>\\nmasterpiece"
533
+ },
534
+ "ADetailer dilate erode": "4",
535
+ "ADetailer mask blur 3rd": "4",
536
+ "ADetailer confidence 3rd": "0.82",
537
+ "ADetailer inpaint padding": "32",
538
+ "ADetailer dilate erode 3rd": "4",
539
+ "ADetailer denoising strength": "0.4",
540
+ "ADetailer use separate steps": "True",
541
+ "ADetailer inpaint only masked": "True",
542
+ "ADetailer inpaint padding 3rd": "32",
543
+ "ADetailer use separate sampler": "True",
544
+ "ADetailer denoising strength 3rd": "0.4",
545
+ "ADetailer inpaint only masked 3rd": "True"
546
+ },
547
+ "availability": "Public",
548
+ "hasMeta": true,
549
+ "hasPositivePrompt": true,
550
+ "onSite": false,
551
+ "remixOfId": null
552
+ },
553
+ {
554
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e9655835-cb61-4190-a676-2c1bd3488f58/original=true/94763283.jpeg",
555
+ "nsfwLevel": 4,
556
+ "width": 1280,
557
+ "height": 1920,
558
+ "hash": "UDHnB8D*00%Lt0o#9vs*02NKSj-S9cEM~BWq",
559
+ "type": "image",
560
+ "metadata": {
561
+ "hash": "UDHnB8D*00%Lt0o#9vs*02NKSj-S9cEM~BWq",
562
+ "size": 2202711,
563
+ "width": 1280,
564
+ "height": 1920
565
+ },
566
+ "minor": false,
567
+ "poi": false,
568
+ "meta": {
569
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
570
+ "Size": "640x960",
571
+ "seed": 3969440177,
572
+ "Model": "spiritforeseermix_spiritforeseerO",
573
+ "steps": 40,
574
+ "hashes": {
575
+ "vae": "f921fb3f29",
576
+ "model": "1e0c69b67c",
577
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
578
+ },
579
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, solo, solo, playboy bunny, fishnets, fake animal ears, rabbit ears, bar \\(place\\), indoors, depth of field, cowboy shot, smirk",
580
+ "Version": "v1.10.1",
581
+ "sampler": "Euler a",
582
+ "cfgScale": 7,
583
+ "resources": [
584
+ {
585
+ "hash": "0x4eb7db2367",
586
+ "name": "Char-IS-Cecilia-V1-SD1.5",
587
+ "type": "lora"
588
+ },
589
+ {
590
+ "hash": "1e0c69b67c",
591
+ "name": "spiritforeseermix_spiritforeseerO",
592
+ "type": "model"
593
+ }
594
+ ],
595
+ "Model hash": "1e0c69b67c",
596
+ "Hires steps": "30",
597
+ "Hires upscale": "2",
598
+ "Schedule type": "Karras",
599
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
600
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
601
+ "ADetailer model": "face_yolov8n.pt",
602
+ "ADetailer steps": "90",
603
+ "ADetailer sampler": "Euler a",
604
+ "ADetailer version": "25.3.0",
605
+ "Denoising strength": "0.35",
606
+ "ADetailer mask blur": "4",
607
+ "ADetailer model 3rd": "hand_yolov8n.pt",
608
+ "ADetailer scheduler": "SGM Uniform",
609
+ "ADetailer confidence": "0.3",
610
+ "ADetailer prompt 3rd": {
611
+ "GoodHands-beta2": "1>\\nmasterpiece"
612
+ },
613
+ "ADetailer dilate erode": "4",
614
+ "ADetailer mask blur 3rd": "4",
615
+ "ADetailer confidence 3rd": "0.82",
616
+ "ADetailer inpaint padding": "32",
617
+ "ADetailer dilate erode 3rd": "4",
618
+ "ADetailer denoising strength": "0.4",
619
+ "ADetailer use separate steps": "True",
620
+ "ADetailer inpaint only masked": "True",
621
+ "ADetailer inpaint padding 3rd": "32",
622
+ "ADetailer use separate sampler": "True",
623
+ "ADetailer denoising strength 3rd": "0.4",
624
+ "ADetailer inpaint only masked 3rd": "True"
625
+ },
626
+ "availability": "Public",
627
+ "hasMeta": true,
628
+ "hasPositivePrompt": true,
629
+ "onSite": false,
630
+ "remixOfId": null
631
+ },
632
+ {
633
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/52a65ea7-b3d7-452f-be7c-8ddc4afae87e/original=true/94763284.jpeg",
634
+ "nsfwLevel": 1,
635
+ "width": 1280,
636
+ "height": 1920,
637
+ "hash": "UQI;ee~V0KNG$#%2tRs:4:NHNIR*W=M|D*Rk",
638
+ "type": "image",
639
+ "metadata": {
640
+ "hash": "UQI;ee~V0KNG$#%2tRs:4:NHNIR*W=M|D*Rk",
641
+ "size": 2295159,
642
+ "width": 1280,
643
+ "height": 1920
644
+ },
645
+ "minor": false,
646
+ "poi": false,
647
+ "meta": {
648
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
649
+ "Size": "640x960",
650
+ "seed": 1780453294,
651
+ "Model": "spiritforeseermix_spiritforeseerO",
652
+ "steps": 40,
653
+ "hashes": {
654
+ "vae": "f921fb3f29",
655
+ "model": "1e0c69b67c",
656
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
657
+ },
658
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, solo, maid, maid headdress, maid apron, embarrassed, white pantyhose, open mouth, blush, smile, open mouth, holding plate, looking at viewer, cowboy shot, bar \\(place\\), indoors, depth of field",
659
+ "Version": "v1.10.1",
660
+ "sampler": "Euler a",
661
+ "cfgScale": 7,
662
+ "resources": [
663
+ {
664
+ "hash": "0x4eb7db2367",
665
+ "name": "Char-IS-Cecilia-V1-SD1.5",
666
+ "type": "lora"
667
+ },
668
+ {
669
+ "hash": "1e0c69b67c",
670
+ "name": "spiritforeseermix_spiritforeseerO",
671
+ "type": "model"
672
+ }
673
+ ],
674
+ "Model hash": "1e0c69b67c",
675
+ "Hires steps": "30",
676
+ "Hires upscale": "2",
677
+ "Schedule type": "Karras",
678
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
679
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
680
+ "ADetailer model": "face_yolov8n.pt",
681
+ "ADetailer steps": "90",
682
+ "ADetailer sampler": "Euler a",
683
+ "ADetailer version": "25.3.0",
684
+ "Denoising strength": "0.35",
685
+ "ADetailer mask blur": "4",
686
+ "ADetailer model 3rd": "hand_yolov8n.pt",
687
+ "ADetailer scheduler": "SGM Uniform",
688
+ "ADetailer confidence": "0.3",
689
+ "ADetailer prompt 3rd": {
690
+ "GoodHands-beta2": "1>\\nmasterpiece"
691
+ },
692
+ "ADetailer dilate erode": "4",
693
+ "ADetailer mask blur 3rd": "4",
694
+ "ADetailer confidence 3rd": "0.82",
695
+ "ADetailer inpaint padding": "32",
696
+ "ADetailer dilate erode 3rd": "4",
697
+ "ADetailer denoising strength": "0.4",
698
+ "ADetailer use separate steps": "True",
699
+ "ADetailer inpaint only masked": "True",
700
+ "ADetailer inpaint padding 3rd": "32",
701
+ "ADetailer use separate sampler": "True",
702
+ "ADetailer denoising strength 3rd": "0.4",
703
+ "ADetailer inpaint only masked 3rd": "True"
704
+ },
705
+ "availability": "Public",
706
+ "hasMeta": true,
707
+ "hasPositivePrompt": true,
708
+ "onSite": false,
709
+ "remixOfId": null
710
+ },
711
+ {
712
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e49161a6-2469-4696-b67b-a87df35bd12a/original=true/94763449.jpeg",
713
+ "nsfwLevel": 4,
714
+ "width": 1280,
715
+ "height": 1920,
716
+ "hash": "U7K1k0_401?w%~?cEQx]0qElE-X9.mtl^*bH",
717
+ "type": "image",
718
+ "metadata": {
719
+ "hash": "U7K1k0_401?w%~?cEQx]0qElE-X9.mtl^*bH",
720
+ "size": 2748518,
721
+ "width": 1280,
722
+ "height": 1920
723
+ },
724
+ "minor": false,
725
+ "poi": false,
726
+ "meta": {
727
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
728
+ "Size": "640x960",
729
+ "seed": 1132025121,
730
+ "Model": "spiritforeseermix_spiritforeseerO",
731
+ "steps": 40,
732
+ "hashes": {
733
+ "vae": "f921fb3f29",
734
+ "model": "1e0c69b67c",
735
+ "lora:body-multi_sagging": "ccb4a56a4a64",
736
+ "lora:cos-bridal_Lingerie": "6b9a5f252458",
737
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
738
+ },
739
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, covered sagging large breasts, cowboy shot, hotel room, window, cityscape, night, looking at viewer, depth of field, light smile, blush, <lora:cos-bridal_Lingerie:0.8> bridal lingerie, lace trim, see-through, garter straps, veil, <lora:body-multi_sagging:0.6>",
740
+ "Version": "v1.10.1",
741
+ "sampler": "Euler a",
742
+ "cfgScale": 7,
743
+ "resources": [
744
+ {
745
+ "hash": "6b9a5f252458",
746
+ "name": "cos-bridal_Lingerie",
747
+ "type": "lora",
748
+ "weight": 0.8
749
+ },
750
+ {
751
+ "hash": "ccb4a56a4a64",
752
+ "name": "body-multi_sagging",
753
+ "type": "lora",
754
+ "weight": 0.6
755
+ },
756
+ {
757
+ "hash": "0x4eb7db2367",
758
+ "name": "Char-IS-Cecilia-V1-SD1.5",
759
+ "type": "lora"
760
+ },
761
+ {
762
+ "hash": "1e0c69b67c",
763
+ "name": "spiritforeseermix_spiritforeseerO",
764
+ "type": "model"
765
+ }
766
+ ],
767
+ "Model hash": "1e0c69b67c",
768
+ "Hires steps": "30",
769
+ "Hires upscale": "2",
770
+ "Schedule type": "Karras",
771
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
772
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
773
+ "ADetailer model": "face_yolov8n.pt",
774
+ "ADetailer steps": "90",
775
+ "ADetailer sampler": "Euler a",
776
+ "ADetailer version": "25.3.0",
777
+ "Denoising strength": "0.35",
778
+ "ADetailer mask blur": "4",
779
+ "ADetailer model 3rd": "hand_yolov8n.pt",
780
+ "ADetailer scheduler": "SGM Uniform",
781
+ "ADetailer confidence": "0.3",
782
+ "ADetailer prompt 3rd": {
783
+ "GoodHands-beta2": "1>\\nmasterpiece"
784
+ },
785
+ "ADetailer dilate erode": "4",
786
+ "ADetailer mask blur 3rd": "4",
787
+ "ADetailer confidence 3rd": "0.82",
788
+ "ADetailer inpaint padding": "32",
789
+ "ADetailer dilate erode 3rd": "4",
790
+ "ADetailer denoising strength": "0.4",
791
+ "ADetailer use separate steps": "True",
792
+ "ADetailer inpaint only masked": "True",
793
+ "ADetailer inpaint padding 3rd": "32",
794
+ "ADetailer use separate sampler": "True",
795
+ "ADetailer denoising strength 3rd": "0.4",
796
+ "ADetailer inpaint only masked 3rd": "True"
797
+ },
798
+ "availability": "Public",
799
+ "hasMeta": true,
800
+ "hasPositivePrompt": true,
801
+ "onSite": false,
802
+ "remixOfId": null
803
+ },
804
+ {
805
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/47d7f05c-d188-4d9d-92ab-6d6b79ee9996/original=true/94763447.jpeg",
806
+ "nsfwLevel": 4,
807
+ "width": 1280,
808
+ "height": 1920,
809
+ "hash": "U5JaD+4n00pJv1Ip0.M{00OF2xjD0zoH|:-p",
810
+ "type": "image",
811
+ "metadata": {
812
+ "hash": "U5JaD+4n00pJv1Ip0.M{00OF2xjD0zoH|:-p",
813
+ "size": 2333601,
814
+ "width": 1280,
815
+ "height": 1920
816
+ },
817
+ "minor": false,
818
+ "poi": false,
819
+ "meta": {
820
+ "VAE": "vae-ft-mse-840000-ema-pruned.ckpt",
821
+ "Size": "640x960",
822
+ "seed": 2258499692,
823
+ "Model": "spiritforeseermix_spiritforeseerO",
824
+ "steps": 40,
825
+ "hashes": {
826
+ "vae": "f921fb3f29",
827
+ "model": "1e0c69b67c",
828
+ "lora:negligee_v0.1": "bb9f6c7b760e",
829
+ "lora:body-multi_sagging": "ccb4a56a4a64",
830
+ "lora:Char-IS-Cecilia-V1-SD1.5": "0x4eb7db2367"
831
+ },
832
+ "prompt": "(best quality, masterpiece:1.2), <lora:Char-IS-Cecilia-V1-SD1.5:0.9:hr=0.6>, 1girl, cecilia alcott, sagging large breasts, <lora:negligee_v0.1:0.8> negligee, cowboy shot, hotel room, window, cityscape, night, looking at viewer, smirk, blush, depth of field, <lora:body-multi_sagging:0.8>",
833
+ "Version": "v1.10.1",
834
+ "sampler": "Euler a",
835
+ "cfgScale": 7,
836
+ "resources": [
837
+ {
838
+ "hash": "bb9f6c7b760e",
839
+ "name": "negligee_v0.1",
840
+ "type": "lora",
841
+ "weight": 0.8
842
+ },
843
+ {
844
+ "hash": "ccb4a56a4a64",
845
+ "name": "body-multi_sagging",
846
+ "type": "lora",
847
+ "weight": 0.8
848
+ },
849
+ {
850
+ "hash": "0x4eb7db2367",
851
+ "name": "Char-IS-Cecilia-V1-SD1.5",
852
+ "type": "lora"
853
+ },
854
+ {
855
+ "hash": "1e0c69b67c",
856
+ "name": "spiritforeseermix_spiritforeseerO",
857
+ "type": "model"
858
+ }
859
+ ],
860
+ "Model hash": "1e0c69b67c",
861
+ "Hires steps": "30",
862
+ "Hires upscale": "2",
863
+ "Schedule type": "Karras",
864
+ "Hires upscaler": "R-ESRGAN 4x+ Anime6B",
865
+ "negativePrompt": "underwear, panties, (worst quality:1.4, low quality:1.4), (bad anatomy), (inaccurate limb:1.2),bad composition, inaccurate eyes, extra digit,fewer digits,(extra arms:1.2), (extra fingers, deformed hands, polydactyl:1.5), badhandv4",
866
+ "ADetailer model": "face_yolov8n.pt",
867
+ "ADetailer steps": "90",
868
+ "ADetailer sampler": "Euler a",
869
+ "ADetailer version": "25.3.0",
870
+ "Denoising strength": "0.35",
871
+ "ADetailer mask blur": "4",
872
+ "ADetailer model 3rd": "hand_yolov8n.pt",
873
+ "ADetailer scheduler": "SGM Uniform",
874
+ "ADetailer confidence": "0.3",
875
+ "ADetailer prompt 3rd": {
876
+ "GoodHands-beta2": "1>\\nmasterpiece"
877
+ },
878
+ "ADetailer dilate erode": "4",
879
+ "ADetailer mask blur 3rd": "4",
880
+ "ADetailer confidence 3rd": "0.82",
881
+ "ADetailer inpaint padding": "32",
882
+ "ADetailer dilate erode 3rd": "4",
883
+ "ADetailer denoising strength": "0.4",
884
+ "ADetailer use separate steps": "True",
885
+ "ADetailer inpaint only masked": "True",
886
+ "ADetailer inpaint padding 3rd": "32",
887
+ "ADetailer use separate sampler": "True",
888
+ "ADetailer denoising strength 3rd": "0.4",
889
+ "ADetailer inpaint only masked 3rd": "True"
890
+ },
891
+ "availability": "Public",
892
+ "hasMeta": true,
893
+ "hasPositivePrompt": true,
894
+ "onSite": false,
895
+ "remixOfId": null
896
+ }
897
+ ],
898
+ "downloadUrl": "https://civitai.com/api/download/models/2121554",
899
+ "creator": {
900
+ "username": "okingjo",
901
+ "image": "https://avatars.githubusercontent.com/u/25336857?v=4"
902
+ },
903
+ "extensions": {
904
+ "sd_civitai_helper": {
905
+ "version": "1.8.13",
906
+ "last_update": 1760877909,
907
+ "skeleton_file": false
908
+ }
909
+ }
910
+ }
A-表番角色/IS〈Infinite Stratos〉/Char-IS-Cecilia-V1-SD1.5.preview.png ADDED

Git LFS Details

  • SHA256: a94d52aa60cab0e71bcd8268b8d5e93adc44732bb16d9cae8c01947f87adb481
  • Pointer size: 132 Bytes
  • Size of remote file: 2.69 MB
A-表番角色/IS〈Infinite Stratos〉/Char-IS-Cecilia-V1-SD1.5.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebb7885f028fc959281d613847c5d51cb8baba18ca2289039fc9c6b501e8ac2e
3
+ size 48108156
A-表番角色/噬血狂袭/15_Aiba_Asagi-30.civitai.info ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2324145,
3
+ "modelId": 2053714,
4
+ "name": "v1.0",
5
+ "nsfwLevel": 60,
6
+ "createdAt": "2025-10-18T03:41:08.996Z",
7
+ "updatedAt": "2025-10-18T03:51:36.704Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-18T03:51:36.678Z",
10
+ "trainedWords": [
11
+ "aiba asagi"
12
+ ],
13
+ "trainingStatus": null,
14
+ "trainingDetails": null,
15
+ "baseModel": "SD 1.5",
16
+ "baseModelType": "Standard",
17
+ "earlyAccessEndsAt": null,
18
+ "earlyAccessConfig": null,
19
+ "description": null,
20
+ "uploadType": "Created",
21
+ "usageControl": "Download",
22
+ "air": "urn:air:sd1:lora:civitai:2053714@2324145",
23
+ "stats": {
24
+ "downloadCount": 35,
25
+ "ratingCount": 0,
26
+ "rating": 0,
27
+ "thumbsUpCount": 12
28
+ },
29
+ "model": {
30
+ "name": "Aiba Asagi | Strike the Blood | \u30b9\u30c8\u30e9\u30a4\u30af\u30fb\u30b6\u30fb\u30d6\u30e9\u30c3\u30c9",
31
+ "type": "LORA",
32
+ "nsfw": true,
33
+ "poi": false,
34
+ "description": "Lora-scale: 0.6-0.8\nTrigger Words: aiba asagi, long hair, brown hair, pink eyes\n\n\n\nIf you'd like to support my work, you can do so through Ko-fi!",
35
+ "tags": [
36
+ "anime",
37
+ "character",
38
+ "strike the blood",
39
+ "girls"
40
+ ],
41
+ "allowNoCredit": true,
42
+ "allowCommercialUse": [
43
+ "RentCivit",
44
+ "Rent"
45
+ ],
46
+ "allowDerivatives": true,
47
+ "allowDifferentLicense": true
48
+ },
49
+ "files": [
50
+ {
51
+ "id": 2214255,
52
+ "sizeKB": 36988.8359375,
53
+ "name": "15_Aiba_Asagi-30.safetensors",
54
+ "type": "Model",
55
+ "pickleScanResult": "Success",
56
+ "pickleScanMessage": "No Pickle imports",
57
+ "virusScanResult": "Success",
58
+ "virusScanMessage": null,
59
+ "scannedAt": "2025-10-18T03:45:38.476Z",
60
+ "metadata": {
61
+ "format": "SafeTensor",
62
+ "size": null,
63
+ "fp": null
64
+ },
65
+ "hashes": {
66
+ "AutoV1": "D0BE9728",
67
+ "AutoV2": "D3D59AC3E4",
68
+ "SHA256": "D3D59AC3E416899B02E9B86CDCABD5FDC7ED17022FFDD9293AA168D914FF867B",
69
+ "CRC32": "A9ADBC28",
70
+ "BLAKE3": "52C8F179865C0737EC6534773D1D44AD2A157C84A3390C097D6827CF376B0294",
71
+ "AutoV3": "39F1C0BE4BE2"
72
+ },
73
+ "primary": true,
74
+ "downloadUrl": "https://civitai.com/api/download/models/2324145"
75
+ }
76
+ ],
77
+ "images": [
78
+ {
79
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/644bcfad-a1d7-4e37-a6b3-af3b3188dea9/original=true/106526314.jpeg",
80
+ "nsfwLevel": 1,
81
+ "width": 512,
82
+ "height": 768,
83
+ "hash": "UBL|4M~V00=|0vaLm;SxTJE2I[sp^lE2Rjjb",
84
+ "type": "image",
85
+ "metadata": {
86
+ "hash": "UBL|4M~V00=|0vaLm;SxTJE2I[sp^lE2Rjjb",
87
+ "size": 445598,
88
+ "width": 512,
89
+ "height": 768,
90
+ "nsfwLevelReason": null
91
+ },
92
+ "minor": false,
93
+ "poi": false,
94
+ "meta": {
95
+ "seed": 462066408449212,
96
+ "Model": "Anyorange1.5",
97
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 462066408449212, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, smile, shirt, indoors, looking at viewer, upper body, pink shirt, short sleeves, ponytail\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Aiba_Asagi-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [462066408449212, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Aiba_Asagi-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, smile, shirt, indoors, looking at viewer, upper body, pink shirt, short sleeves, ponytail\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [336.02875955226216, -26.349484917656344]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
98
+ "steps": 30,
99
+ "width": 512,
100
+ "height": 768,
101
+ "models": [
102
+ "Anyorange1.5.safetensors"
103
+ ],
104
+ "prompt": "masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, smile, shirt, indoors, looking at viewer, upper body, pink shirt, short sleeves, ponytail",
105
+ "denoise": 1,
106
+ "sampler": "Euler a",
107
+ "cfgScale": 5,
108
+ "scheduler": "normal",
109
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
110
+ "additionalResources": [
111
+ {
112
+ "name": "1.5/Aiba_Asagi-30.safetensors",
113
+ "type": "lora",
114
+ "strength": 0.7000000000000002,
115
+ "strengthClip": 1
116
+ }
117
+ ]
118
+ },
119
+ "availability": "Public",
120
+ "hasMeta": true,
121
+ "hasPositivePrompt": true,
122
+ "onSite": false,
123
+ "remixOfId": null
124
+ },
125
+ {
126
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e6471b97-5002-448b-b487-213a3ba4503c/original=true/106526313.jpeg",
127
+ "nsfwLevel": 1,
128
+ "width": 512,
129
+ "height": 768,
130
+ "hash": "UDK1an~V9w~Xxb%N?wx[9ZV@IARjjCf5Mys:",
131
+ "type": "image",
132
+ "metadata": {
133
+ "hash": "UDK1an~V9w~Xxb%N?wx[9ZV@IARjjCf5Mys:",
134
+ "size": 469225,
135
+ "width": 512,
136
+ "height": 768,
137
+ "nsfwLevelReason": null
138
+ },
139
+ "minor": false,
140
+ "poi": false,
141
+ "meta": {
142
+ "seed": 165878887903546,
143
+ "Model": "Anyorange1.5",
144
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 165878887903546, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, open mouth, school uniform, smile, sailor collar, anime coloring, collarbone, :d\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Aiba_Asagi-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [165878887903546, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Aiba_Asagi-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, open mouth, school uniform, smile, sailor collar, anime coloring, collarbone, :d\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [336.02875955226216, -26.349484917656344]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
145
+ "steps": 30,
146
+ "width": 512,
147
+ "height": 768,
148
+ "models": [
149
+ "Anyorange1.5.safetensors"
150
+ ],
151
+ "prompt": "masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, open mouth, school uniform, smile, sailor collar, anime coloring, collarbone, :d",
152
+ "denoise": 1,
153
+ "sampler": "Euler a",
154
+ "cfgScale": 5,
155
+ "scheduler": "normal",
156
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
157
+ "additionalResources": [
158
+ {
159
+ "name": "1.5/Aiba_Asagi-30.safetensors",
160
+ "type": "lora",
161
+ "strength": 0.7000000000000002,
162
+ "strengthClip": 1
163
+ }
164
+ ]
165
+ },
166
+ "availability": "Public",
167
+ "hasMeta": true,
168
+ "hasPositivePrompt": true,
169
+ "onSite": false,
170
+ "remixOfId": null
171
+ },
172
+ {
173
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/1b5d1c54-3db9-4e7d-9da7-78372ac33ada/original=true/106526315.jpeg",
174
+ "nsfwLevel": 1,
175
+ "width": 512,
176
+ "height": 768,
177
+ "hash": "UCJkJ%Vq01Vv00%0tlNb%JNIM#x]~poI-;IV",
178
+ "type": "image",
179
+ "metadata": {
180
+ "hash": "UCJkJ%Vq01Vv00%0tlNb%JNIM#x]~poI-;IV",
181
+ "size": 489679,
182
+ "width": 512,
183
+ "height": 768,
184
+ "nsfwLevelReason": null
185
+ },
186
+ "minor": false,
187
+ "poi": false,
188
+ "meta": {
189
+ "seed": 20398364859764,
190
+ "Model": "Anyorange1.5",
191
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 20398364859764, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, school uniform, ponytail, scrunchie, hair ornament, serafuku, open mouth, upper body\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Aiba_Asagi-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [20398364859764, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Aiba_Asagi-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, school uniform, ponytail, scrunchie, hair ornament, serafuku, open mouth, upper body\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [336.02875955226216, -26.349484917656344]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
192
+ "steps": 30,
193
+ "width": 512,
194
+ "height": 768,
195
+ "models": [
196
+ "Anyorange1.5.safetensors"
197
+ ],
198
+ "prompt": "masterpiece,best quality,amazing quality, aiba asagi, long hair, brown hair, pink eyes, 1girl, solo, school uniform, ponytail, scrunchie, hair ornament, serafuku, open mouth, upper body",
199
+ "denoise": 1,
200
+ "sampler": "Euler a",
201
+ "cfgScale": 5,
202
+ "scheduler": "normal",
203
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
204
+ "additionalResources": [
205
+ {
206
+ "name": "1.5/Aiba_Asagi-30.safetensors",
207
+ "type": "lora",
208
+ "strength": 0.7000000000000002,
209
+ "strengthClip": 1
210
+ }
211
+ ]
212
+ },
213
+ "availability": "Public",
214
+ "hasMeta": true,
215
+ "hasPositivePrompt": true,
216
+ "onSite": false,
217
+ "remixOfId": null
218
+ }
219
+ ],
220
+ "downloadUrl": "https://civitai.com/api/download/models/2324145",
221
+ "creator": {
222
+ "username": "duongve13112002",
223
+ "image": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f5b736ce-ca77-4637-896d-475236c4196e/width=96/duongve13112002.jpeg"
224
+ },
225
+ "extensions": {
226
+ "sd_civitai_helper": {
227
+ "version": "1.8.13",
228
+ "last_update": 1760877705,
229
+ "skeleton_file": false
230
+ }
231
+ }
232
+ }
A-表番角色/噬血狂袭/15_Aiba_Asagi-30.preview.png ADDED

Git LFS Details

  • SHA256: 5b3e0b7b90020d33f568b4c1213f239ea7f96f90157a0664aa71e89f6adb12d9
  • Pointer size: 131 Bytes
  • Size of remote file: 446 kB
A-表番角色/噬血狂袭/15_Aiba_Asagi-30.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d3d59ac3e416899b02e9b86cdcabd5fdc7ed17022ffdd9293aa168d914ff867b
3
+ size 37876568
A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.civitai.info ADDED
@@ -0,0 +1,185 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2324174,
3
+ "modelId": 2053741,
4
+ "name": "v1.0",
5
+ "nsfwLevel": 60,
6
+ "createdAt": "2025-10-18T03:58:04.425Z",
7
+ "updatedAt": "2025-10-18T04:04:30.408Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-18T04:04:30.374Z",
10
+ "trainedWords": [
11
+ "akatsuki nagisa"
12
+ ],
13
+ "trainingStatus": null,
14
+ "trainingDetails": null,
15
+ "baseModel": "SD 1.5",
16
+ "baseModelType": "Standard",
17
+ "earlyAccessEndsAt": null,
18
+ "earlyAccessConfig": null,
19
+ "description": null,
20
+ "uploadType": "Created",
21
+ "usageControl": "Download",
22
+ "air": "urn:air:sd1:lora:civitai:2053741@2324174",
23
+ "stats": {
24
+ "downloadCount": 36,
25
+ "ratingCount": 0,
26
+ "rating": 0,
27
+ "thumbsUpCount": 12
28
+ },
29
+ "model": {
30
+ "name": "Akatsuki Nagisa | Strike the Blood | \u30b9\u30c8\u30e9\u30a4\u30af\u30fb\u30b6\u30fb\u30d6\u30e9\u30c3\u30c9",
31
+ "type": "LORA",
32
+ "nsfw": true,
33
+ "poi": false,
34
+ "description": "Lora-scale: 0.6-0.8\nTrigger Words: akatsuki nagisa, long hair, black hair, red eyes, ponytail\n\n\n\nIf you'd like to support my work, you can do so through Ko-fi!",
35
+ "tags": [
36
+ "anime",
37
+ "character",
38
+ "strike the blood",
39
+ "girls"
40
+ ],
41
+ "allowNoCredit": true,
42
+ "allowCommercialUse": [
43
+ "RentCivit",
44
+ "Rent"
45
+ ],
46
+ "allowDerivatives": true,
47
+ "allowDifferentLicense": true
48
+ },
49
+ "files": [
50
+ {
51
+ "id": 2214291,
52
+ "sizeKB": 36981.703125,
53
+ "name": "15_Akatsuki_Nagisa-30.safetensors",
54
+ "type": "Model",
55
+ "pickleScanResult": "Success",
56
+ "pickleScanMessage": "No Pickle imports",
57
+ "virusScanResult": "Success",
58
+ "virusScanMessage": null,
59
+ "scannedAt": "2025-10-18T04:00:32.598Z",
60
+ "metadata": {
61
+ "format": "SafeTensor",
62
+ "size": null,
63
+ "fp": null
64
+ },
65
+ "hashes": {
66
+ "AutoV1": "9544B8AD",
67
+ "AutoV2": "30F1E6BE30",
68
+ "SHA256": "30F1E6BE3050A6653B41CD03AA122ED3E9D81127D429233D081DE3FBFCBA6DAA",
69
+ "CRC32": "9A372CBD",
70
+ "BLAKE3": "85E2B3B92E8665703665FD1484A70FAE1FD8D1B6CDD6D32C0AC08C98D743381A",
71
+ "AutoV3": "167FF006EA08"
72
+ },
73
+ "primary": true,
74
+ "downloadUrl": "https://civitai.com/api/download/models/2324174"
75
+ }
76
+ ],
77
+ "images": [
78
+ {
79
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/87c65e0a-1a48-4296-a78b-52f45ed28dbe/original=true/106527544.jpeg",
80
+ "nsfwLevel": 1,
81
+ "width": 512,
82
+ "height": 768,
83
+ "hash": "UBHU,s4;8yE102f%M|sp}uRQTut55PIo-pj@",
84
+ "type": "image",
85
+ "metadata": {
86
+ "hash": "UBHU,s4;8yE102f%M|sp}uRQTut55PIo-pj@",
87
+ "size": 433366,
88
+ "width": 512,
89
+ "height": 768,
90
+ "nsfwLevelReason": null
91
+ },
92
+ "minor": false,
93
+ "poi": false,
94
+ "meta": {
95
+ "seed": 61453113666858,
96
+ "Model": "Anyorange1.5",
97
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 61453113666858, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, akatsuki nagisa, long hair, black hair, red eyes, ponytail, 1girl, solo, dress, smile, lying, looking at viewer, bow, on back, arms up, necklace, hair bow, green dress, ribbon, short sleeves\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Akatsuki_Nagisa-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [61453113666858, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Akatsuki_Nagisa-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, akatsuki nagisa, long hair, black hair, red eyes, ponytail, 1girl, solo, dress, smile, lying, looking at viewer, bow, on back, arms up, necklace, hair bow, green dress, ribbon, short sleeves\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [129.73594020445427, -55.81988177718275]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
98
+ "steps": 30,
99
+ "width": 512,
100
+ "height": 768,
101
+ "models": [
102
+ "Anyorange1.5.safetensors"
103
+ ],
104
+ "prompt": "masterpiece,best quality,amazing quality, akatsuki nagisa, long hair, black hair, red eyes, ponytail, 1girl, solo, dress, smile, lying, looking at viewer, bow, on back, arms up, necklace, hair bow, green dress, ribbon, short sleeves",
105
+ "denoise": 1,
106
+ "sampler": "Euler a",
107
+ "cfgScale": 5,
108
+ "scheduler": "normal",
109
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
110
+ "additionalResources": [
111
+ {
112
+ "name": "1.5/Akatsuki_Nagisa-30.safetensors",
113
+ "type": "lora",
114
+ "strength": 0.7000000000000002,
115
+ "strengthClip": 1
116
+ }
117
+ ]
118
+ },
119
+ "availability": "Public",
120
+ "hasMeta": true,
121
+ "hasPositivePrompt": true,
122
+ "onSite": false,
123
+ "remixOfId": null
124
+ },
125
+ {
126
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/7cb398e6-e02c-4581-aaf5-d81aeebee214/original=true/106527542.jpeg",
127
+ "nsfwLevel": 1,
128
+ "width": 512,
129
+ "height": 768,
130
+ "hash": "UHK-Xuxu^$-;DhE1_NoLxD-:s:xuT1t7-TD*",
131
+ "type": "image",
132
+ "metadata": {
133
+ "hash": "UHK-Xuxu^$-;DhE1_NoLxD-:s:xuT1t7-TD*",
134
+ "size": 401246,
135
+ "width": 512,
136
+ "height": 768,
137
+ "nsfwLevelReason": null
138
+ },
139
+ "minor": false,
140
+ "poi": false,
141
+ "meta": {
142
+ "seed": 1064018633082876,
143
+ "Model": "Anyorange1.5",
144
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 1064018633082876, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, akatsuki nagisa, long hair, black hair, red eyes, ponytail, 1girl, solo, school uniform, bow, serafuku, looking at viewer, upper body, smile, shirt, sailor collar, hair bow, white shirt, blue sailor collar, indoors\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Akatsuki_Nagisa-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [1064018633082876, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Akatsuki_Nagisa-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, akatsuki nagisa, long hair, black hair, red eyes, ponytail, 1girl, solo, school uniform, bow, serafuku, looking at viewer, upper body, smile, shirt, sailor collar, hair bow, white shirt, blue sailor collar, indoors\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [20.522100988609996, -26.349468385207096]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
145
+ "steps": 30,
146
+ "width": 512,
147
+ "height": 768,
148
+ "models": [
149
+ "Anyorange1.5.safetensors"
150
+ ],
151
+ "prompt": "masterpiece,best quality,amazing quality, akatsuki nagisa, long hair, black hair, red eyes, ponytail, 1girl, solo, school uniform, bow, serafuku, looking at viewer, upper body, smile, shirt, sailor collar, hair bow, white shirt, blue sailor collar, indoors",
152
+ "denoise": 1,
153
+ "sampler": "Euler a",
154
+ "cfgScale": 5,
155
+ "scheduler": "normal",
156
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
157
+ "additionalResources": [
158
+ {
159
+ "name": "1.5/Akatsuki_Nagisa-30.safetensors",
160
+ "type": "lora",
161
+ "strength": 0.7000000000000002,
162
+ "strengthClip": 1
163
+ }
164
+ ]
165
+ },
166
+ "availability": "Public",
167
+ "hasMeta": true,
168
+ "hasPositivePrompt": true,
169
+ "onSite": false,
170
+ "remixOfId": null
171
+ }
172
+ ],
173
+ "downloadUrl": "https://civitai.com/api/download/models/2324174",
174
+ "creator": {
175
+ "username": "duongve13112002",
176
+ "image": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f5b736ce-ca77-4637-896d-475236c4196e/width=96/duongve13112002.jpeg"
177
+ },
178
+ "extensions": {
179
+ "sd_civitai_helper": {
180
+ "version": "1.8.13",
181
+ "last_update": 1760877711,
182
+ "skeleton_file": false
183
+ }
184
+ }
185
+ }
A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.preview.png ADDED

Git LFS Details

  • SHA256: fdf015279f66fb0819b89771d47411c20479a8be85b03e7cba37337edb220da3
  • Pointer size: 131 Bytes
  • Size of remote file: 433 kB
A-表番角色/噬血狂袭/15_Akatsuki_Nagisa-30.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:30f1e6be3050a6653b41cd03aa122ed3e9d81127d429233d081de3fbfcba6daa
3
+ size 37869264
A-表番角色/噬血狂袭/15_Astarte-30.civitai.info ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2326808,
3
+ "modelId": 2056117,
4
+ "name": "v1.0",
5
+ "nsfwLevel": 60,
6
+ "createdAt": "2025-10-19T00:28:46.339Z",
7
+ "updatedAt": "2025-10-19T00:32:42.114Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-19T00:32:42.085Z",
10
+ "trainedWords": [
11
+ "astarte (strike the blood)"
12
+ ],
13
+ "trainingStatus": null,
14
+ "trainingDetails": null,
15
+ "baseModel": "SD 1.5",
16
+ "baseModelType": "Standard",
17
+ "earlyAccessEndsAt": null,
18
+ "earlyAccessConfig": null,
19
+ "description": null,
20
+ "uploadType": "Created",
21
+ "usageControl": "Download",
22
+ "air": "urn:air:sd1:lora:civitai:2056117@2326808",
23
+ "stats": {
24
+ "downloadCount": 19,
25
+ "ratingCount": 0,
26
+ "rating": 0,
27
+ "thumbsUpCount": 9
28
+ },
29
+ "model": {
30
+ "name": "Astarte | Strike the Blood | \u30b9\u30c8\u30e9\u30a4\u30af\u30fb\u30b6\u30fb\u30d6\u30e9\u30c3\u30c9",
31
+ "type": "LORA",
32
+ "nsfw": true,
33
+ "poi": false,
34
+ "description": "Lora-scale: 0.6-0.8\nTrigger Words: astarte (strike the blood), blue hair, blue eyes, long hair\n\n\n\nIf you'd like to support my work, you can do so through Ko-fi!",
35
+ "tags": [
36
+ "anime",
37
+ "character",
38
+ "strike the blood",
39
+ "girls"
40
+ ],
41
+ "allowNoCredit": true,
42
+ "allowCommercialUse": [
43
+ "RentCivit",
44
+ "Rent"
45
+ ],
46
+ "allowDerivatives": true,
47
+ "allowDifferentLicense": true
48
+ },
49
+ "files": [
50
+ {
51
+ "id": 2217016,
52
+ "sizeKB": 36982.1875,
53
+ "name": "15_Astarte-30.safetensors",
54
+ "type": "Model",
55
+ "pickleScanResult": "Success",
56
+ "pickleScanMessage": "No Pickle imports",
57
+ "virusScanResult": "Success",
58
+ "virusScanMessage": null,
59
+ "scannedAt": "2025-10-19T00:30:30.688Z",
60
+ "metadata": {
61
+ "format": "SafeTensor",
62
+ "size": null,
63
+ "fp": null
64
+ },
65
+ "hashes": {
66
+ "AutoV1": "14FCA386",
67
+ "AutoV2": "77691173A1",
68
+ "SHA256": "77691173A1D2D97ED1724B4FF51DC5237F5D6C5DE582EF66EBEF976F20FF7CD4",
69
+ "CRC32": "7EA43D97",
70
+ "BLAKE3": "E899A20D044D737878D31DA4AADB307FE6DA29AF49DD8B3D6A709B6E3D8FE4F6",
71
+ "AutoV3": "C19CB0689E19"
72
+ },
73
+ "primary": true,
74
+ "downloadUrl": "https://civitai.com/api/download/models/2326808"
75
+ }
76
+ ],
77
+ "images": [
78
+ {
79
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f832f537-6600-4751-9007-9291c15ce070/original=true/106645071.jpeg",
80
+ "nsfwLevel": 1,
81
+ "width": 512,
82
+ "height": 768,
83
+ "hash": "UHJa.t_NND9GT1%M~p%M%gM{-:V@oht7IUjY",
84
+ "type": "image",
85
+ "metadata": {
86
+ "hash": "UHJa.t_NND9GT1%M~p%M%gM{-:V@oht7IUjY",
87
+ "size": 481716,
88
+ "width": 512,
89
+ "height": 768,
90
+ "nsfwLevelReason": null
91
+ },
92
+ "minor": false,
93
+ "poi": false,
94
+ "meta": {
95
+ "seed": 784041803250288,
96
+ "Model": "Anyorange1.5",
97
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 784041803250288, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, anime coloring, indoors, maid, maid headdress, ribbon, hairband, hair between eyes, red ribbon\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Astarte-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [784041803250288, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Astarte-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, anime coloring, indoors, maid, maid headdress, ribbon, hairband, hair between eyes, red ribbon\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [371.5665838871727, -28.082961819331874]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
98
+ "steps": 30,
99
+ "width": 512,
100
+ "height": 768,
101
+ "models": [
102
+ "Anyorange1.5.safetensors"
103
+ ],
104
+ "prompt": "masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, anime coloring, indoors, maid, maid headdress, ribbon, hairband, hair between eyes, red ribbon",
105
+ "denoise": 1,
106
+ "sampler": "Euler a",
107
+ "cfgScale": 5,
108
+ "scheduler": "normal",
109
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
110
+ "additionalResources": [
111
+ {
112
+ "name": "1.5/Astarte-30.safetensors",
113
+ "type": "lora",
114
+ "strength": 0.7000000000000002,
115
+ "strengthClip": 1
116
+ }
117
+ ]
118
+ },
119
+ "availability": "Public",
120
+ "hasMeta": true,
121
+ "hasPositivePrompt": true,
122
+ "onSite": false,
123
+ "remixOfId": null
124
+ },
125
+ {
126
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/a9059da9-3556-46c4-a276-3de6dfb1062f/original=true/106645072.jpeg",
127
+ "nsfwLevel": 1,
128
+ "width": 512,
129
+ "height": 768,
130
+ "hash": "UHI=AU~qEJIV0Mo#-:xa_NtQ-:%M_3-:t7V@",
131
+ "type": "image",
132
+ "metadata": {
133
+ "hash": "UHI=AU~qEJIV0Mo#-:xa_NtQ-:%M_3-:t7V@",
134
+ "size": 499373,
135
+ "width": 512,
136
+ "height": 768,
137
+ "nsfwLevelReason": null
138
+ },
139
+ "minor": false,
140
+ "poi": false,
141
+ "meta": {
142
+ "seed": 634755387290929,
143
+ "Model": "Anyorange1.5",
144
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 634755387290929, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, maid, red ribbon, hair between eyes, ribbon, maid headdress, looking at viewer, short sleeves, neck ribbon, very long hair, apron, shiny hair, white apron\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Astarte-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [634755387290929, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Astarte-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, maid, red ribbon, hair between eyes, ribbon, maid headdress, looking at viewer, short sleeves, neck ribbon, very long hair, apron, shiny hair, white apron\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [371.5665838871727, -28.082961819331874]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
145
+ "steps": 30,
146
+ "width": 512,
147
+ "height": 768,
148
+ "models": [
149
+ "Anyorange1.5.safetensors"
150
+ ],
151
+ "prompt": "masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, maid, red ribbon, hair between eyes, ribbon, maid headdress, looking at viewer, short sleeves, neck ribbon, very long hair, apron, shiny hair, white apron",
152
+ "denoise": 1,
153
+ "sampler": "Euler a",
154
+ "cfgScale": 5,
155
+ "scheduler": "normal",
156
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
157
+ "additionalResources": [
158
+ {
159
+ "name": "1.5/Astarte-30.safetensors",
160
+ "type": "lora",
161
+ "strength": 0.7000000000000002,
162
+ "strengthClip": 1
163
+ }
164
+ ]
165
+ },
166
+ "availability": "Public",
167
+ "hasMeta": true,
168
+ "hasPositivePrompt": true,
169
+ "onSite": false,
170
+ "remixOfId": null
171
+ },
172
+ {
173
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9df0cb38-287a-4631-bf04-1f08abad91af/original=true/106645073.jpeg",
174
+ "nsfwLevel": 1,
175
+ "width": 512,
176
+ "height": 768,
177
+ "hash": "UIK-5@I8?]-TO[I9-Oxs?wbc^*%Lo#ogIn%L",
178
+ "type": "image",
179
+ "metadata": {
180
+ "hash": "UIK-5@I8?]-TO[I9-Oxs?wbc^*%Lo#ogIn%L",
181
+ "size": 459724,
182
+ "width": 512,
183
+ "height": 768,
184
+ "nsfwLevelReason": null
185
+ },
186
+ "minor": false,
187
+ "poi": false,
188
+ "meta": {
189
+ "seed": 595554708277818,
190
+ "Model": "Anyorange1.5",
191
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 595554708277818, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, looking at viewer, pajamas, shirt\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Astarte-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [595554708277818, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Astarte-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, looking at viewer, pajamas, shirt\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [371.5665838871727, -28.082961819331874]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
192
+ "steps": 30,
193
+ "width": 512,
194
+ "height": 768,
195
+ "models": [
196
+ "Anyorange1.5.safetensors"
197
+ ],
198
+ "prompt": "masterpiece,best quality,amazing quality, astarte (strike the blood), blue hair, blue eyes, long hair, 1girl, solo, looking at viewer, pajamas, shirt",
199
+ "denoise": 1,
200
+ "sampler": "Euler a",
201
+ "cfgScale": 5,
202
+ "scheduler": "normal",
203
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
204
+ "additionalResources": [
205
+ {
206
+ "name": "1.5/Astarte-30.safetensors",
207
+ "type": "lora",
208
+ "strength": 0.7000000000000002,
209
+ "strengthClip": 1
210
+ }
211
+ ]
212
+ },
213
+ "availability": "Public",
214
+ "hasMeta": true,
215
+ "hasPositivePrompt": true,
216
+ "onSite": false,
217
+ "remixOfId": null
218
+ }
219
+ ],
220
+ "downloadUrl": "https://civitai.com/api/download/models/2326808",
221
+ "creator": {
222
+ "username": "duongve13112002",
223
+ "image": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f5b736ce-ca77-4637-896d-475236c4196e/width=96/duongve13112002.jpeg"
224
+ },
225
+ "extensions": {
226
+ "sd_civitai_helper": {
227
+ "version": "1.8.13",
228
+ "last_update": 1760877715,
229
+ "skeleton_file": false
230
+ }
231
+ }
232
+ }
A-表番角色/噬血狂袭/15_Astarte-30.preview.png ADDED

Git LFS Details

  • SHA256: a4f85cd1be89627c017352273b6872476850ef306f2c04c67702d18c72e5bec9
  • Pointer size: 131 Bytes
  • Size of remote file: 482 kB
A-表番角色/噬血狂袭/15_Astarte-30.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77691173a1d2d97ed1724b4ff51dc5237f5d6c5de582ef66ebef976f20ff7cd4
3
+ size 37869760
A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.civitai.info ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2326828,
3
+ "modelId": 2056137,
4
+ "name": "v1.0",
5
+ "nsfwLevel": 60,
6
+ "createdAt": "2025-10-19T00:38:12.512Z",
7
+ "updatedAt": "2025-10-19T00:43:39.145Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-19T00:43:39.086Z",
10
+ "trainedWords": [
11
+ "himeragi yukina"
12
+ ],
13
+ "trainingStatus": null,
14
+ "trainingDetails": null,
15
+ "baseModel": "SD 1.5",
16
+ "baseModelType": "Standard",
17
+ "earlyAccessEndsAt": null,
18
+ "earlyAccessConfig": null,
19
+ "description": null,
20
+ "uploadType": "Created",
21
+ "usageControl": "Download",
22
+ "air": "urn:air:sd1:lora:civitai:2056137@2326828",
23
+ "stats": {
24
+ "downloadCount": 23,
25
+ "ratingCount": 0,
26
+ "rating": 0,
27
+ "thumbsUpCount": 9
28
+ },
29
+ "model": {
30
+ "name": "Himeragi Yukina | Strike the Blood | \u30b9\u30c8\u30e9\u30a4\u30af\u30fb\u30b6\u30fb\u30d6\u30e9\u30c3\u30c9",
31
+ "type": "LORA",
32
+ "nsfw": true,
33
+ "poi": false,
34
+ "description": "Lora-scale: 0.6-0.8\nTrigger Words: himeragi yukina, long hair, brown eyes, black hair\n\n\n\nIf you'd like to support my work, you can do so through Ko-fi!",
35
+ "tags": [
36
+ "anime",
37
+ "character",
38
+ "strike the blood",
39
+ "girls"
40
+ ],
41
+ "allowNoCredit": true,
42
+ "allowCommercialUse": [
43
+ "RentCivit",
44
+ "Rent"
45
+ ],
46
+ "allowDerivatives": true,
47
+ "allowDifferentLicense": true
48
+ },
49
+ "files": [
50
+ {
51
+ "id": 2217033,
52
+ "sizeKB": 36990.0859375,
53
+ "name": "15_Himeragi_Yukina-29.safetensors",
54
+ "type": "Model",
55
+ "pickleScanResult": "Success",
56
+ "pickleScanMessage": "No Pickle imports",
57
+ "virusScanResult": "Success",
58
+ "virusScanMessage": null,
59
+ "scannedAt": "2025-10-19T00:40:44.843Z",
60
+ "metadata": {
61
+ "format": "SafeTensor",
62
+ "size": null,
63
+ "fp": null
64
+ },
65
+ "hashes": {
66
+ "AutoV1": "9837D54F",
67
+ "AutoV2": "0B6B150720",
68
+ "SHA256": "0B6B15072053F101346747650A9D989AECB167DB4E19D45CA57CD799FDC52EC2",
69
+ "CRC32": "C6D79E6C",
70
+ "BLAKE3": "CE77CAB8147A1D6C3D221F596F014C97240AEBF3E4ADD41D495E95392C69C2BE",
71
+ "AutoV3": "8D3FEBD58075"
72
+ },
73
+ "primary": true,
74
+ "downloadUrl": "https://civitai.com/api/download/models/2326828"
75
+ }
76
+ ],
77
+ "images": [
78
+ {
79
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/54eee640-2b2e-4408-98d9-1632423320c1/original=true/106646213.jpeg",
80
+ "nsfwLevel": 1,
81
+ "width": 512,
82
+ "height": 768,
83
+ "hash": "U9F=js?aH;?a00?aNLt8=pjF~qIp01R,xaM{",
84
+ "type": "image",
85
+ "metadata": {
86
+ "hash": "U9F=js?aH;?a00?aNLt8=pjF~qIp01R,xaM{",
87
+ "size": 401400,
88
+ "width": 512,
89
+ "height": 768,
90
+ "nsfwLevelReason": null
91
+ },
92
+ "minor": false,
93
+ "poi": false,
94
+ "meta": {
95
+ "seed": 679012508623869,
96
+ "Model": "Anyorange1.5",
97
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 679012508623869, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, school uniform, indoors, serafuku, blue sailor collar, short sleeves, shirt, white shirt, sailor collar, bow\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Himeragi_Yukina-29.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [679012508623869, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Himeragi_Yukina-29.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, school uniform, indoors, serafuku, blue sailor collar, short sleeves, shirt, white shirt, sailor collar, bow\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [149.67180615366368, 3.1210276690148713]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
98
+ "steps": 30,
99
+ "width": 512,
100
+ "height": 768,
101
+ "models": [
102
+ "Anyorange1.5.safetensors"
103
+ ],
104
+ "prompt": "masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, school uniform, indoors, serafuku, blue sailor collar, short sleeves, shirt, white shirt, sailor collar, bow",
105
+ "denoise": 1,
106
+ "sampler": "Euler a",
107
+ "cfgScale": 5,
108
+ "scheduler": "normal",
109
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
110
+ "additionalResources": [
111
+ {
112
+ "name": "1.5/Himeragi_Yukina-29.safetensors",
113
+ "type": "lora",
114
+ "strength": 0.7000000000000002,
115
+ "strengthClip": 1
116
+ }
117
+ ]
118
+ },
119
+ "availability": "Public",
120
+ "hasMeta": true,
121
+ "hasPositivePrompt": true,
122
+ "onSite": false,
123
+ "remixOfId": null
124
+ },
125
+ {
126
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f3e685c4-6d58-4188-8bbb-ae64968c921c/original=true/106646214.jpeg",
127
+ "nsfwLevel": 1,
128
+ "width": 512,
129
+ "height": 768,
130
+ "hash": "UJG8l%NJ~T^*4:t7IpRkxqoIxHfR0LWAIUM{",
131
+ "type": "image",
132
+ "metadata": {
133
+ "hash": "UJG8l%NJ~T^*4:t7IpRkxqoIxHfR0LWAIUM{",
134
+ "size": 465309,
135
+ "width": 512,
136
+ "height": 768,
137
+ "nsfwLevelReason": null
138
+ },
139
+ "minor": false,
140
+ "poi": false,
141
+ "meta": {
142
+ "seed": 86802468711994,
143
+ "Model": "Anyorange1.5",
144
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 86802468711994, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, japanese clothes, kimono, blue kimono, instrument case, guitar case, obi, sash\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Himeragi_Yukina-29.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [86802468711994, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Himeragi_Yukina-29.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, japanese clothes, kimono, blue kimono, instrument case, guitar case, obi, sash\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [149.67180615366368, 3.1210276690148713]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
145
+ "steps": 30,
146
+ "width": 512,
147
+ "height": 768,
148
+ "models": [
149
+ "Anyorange1.5.safetensors"
150
+ ],
151
+ "prompt": "masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, japanese clothes, kimono, blue kimono, instrument case, guitar case, obi, sash",
152
+ "denoise": 1,
153
+ "sampler": "Euler a",
154
+ "cfgScale": 5,
155
+ "scheduler": "normal",
156
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
157
+ "additionalResources": [
158
+ {
159
+ "name": "1.5/Himeragi_Yukina-29.safetensors",
160
+ "type": "lora",
161
+ "strength": 0.7000000000000002,
162
+ "strengthClip": 1
163
+ }
164
+ ]
165
+ },
166
+ "availability": "Public",
167
+ "hasMeta": true,
168
+ "hasPositivePrompt": true,
169
+ "onSite": false,
170
+ "remixOfId": null
171
+ },
172
+ {
173
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/29c31195-1630-42e5-b5fa-89fe5c3637dd/original=true/106646215.jpeg",
174
+ "nsfwLevel": 1,
175
+ "width": 512,
176
+ "height": 768,
177
+ "hash": "U8HnvsI800-p00?H~XRkH=M{%#M{?bxt9a9#",
178
+ "type": "image",
179
+ "metadata": {
180
+ "hash": "U8HnvsI800-p00?H~XRkH=M{%#M{?bxt9a9#",
181
+ "size": 448692,
182
+ "width": 512,
183
+ "height": 768,
184
+ "nsfwLevelReason": null
185
+ },
186
+ "minor": false,
187
+ "poi": false,
188
+ "meta": {
189
+ "seed": 298422980031485,
190
+ "Model": "Anyorange1.5",
191
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 298422980031485, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, school uniform, shirt, white shirt, indoors, blue sailor collar, hair between eyes, serafuku, sailor collar, bookshelf, bow, upper body, short sleeves\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Himeragi_Yukina-29.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [298422980031485, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Himeragi_Yukina-29.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, school uniform, shirt, white shirt, indoors, blue sailor collar, hair between eyes, serafuku, sailor collar, bookshelf, bow, upper body, short sleeves\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [149.67180615366368, 3.1210276690148713]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
192
+ "steps": 30,
193
+ "width": 512,
194
+ "height": 768,
195
+ "models": [
196
+ "Anyorange1.5.safetensors"
197
+ ],
198
+ "prompt": "masterpiece,best quality,amazing quality, himeragi yukina, long hair, brown eyes, black hair, 1girl, solo, school uniform, shirt, white shirt, indoors, blue sailor collar, hair between eyes, serafuku, sailor collar, bookshelf, bow, upper body, short sleeves",
199
+ "denoise": 1,
200
+ "sampler": "Euler a",
201
+ "cfgScale": 5,
202
+ "scheduler": "normal",
203
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
204
+ "additionalResources": [
205
+ {
206
+ "name": "1.5/Himeragi_Yukina-29.safetensors",
207
+ "type": "lora",
208
+ "strength": 0.7000000000000002,
209
+ "strengthClip": 1
210
+ }
211
+ ]
212
+ },
213
+ "availability": "Public",
214
+ "hasMeta": true,
215
+ "hasPositivePrompt": true,
216
+ "onSite": false,
217
+ "remixOfId": null
218
+ }
219
+ ],
220
+ "downloadUrl": "https://civitai.com/api/download/models/2326828",
221
+ "creator": {
222
+ "username": "duongve13112002",
223
+ "image": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f5b736ce-ca77-4637-896d-475236c4196e/width=96/duongve13112002.jpeg"
224
+ },
225
+ "extensions": {
226
+ "sd_civitai_helper": {
227
+ "version": "1.8.13",
228
+ "last_update": 1760877718,
229
+ "skeleton_file": false
230
+ }
231
+ }
232
+ }
A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.preview.png ADDED

Git LFS Details

  • SHA256: 98fac4e09c196b0b90d403af2394e314ff5883c207c9f502f480d69507cd9c60
  • Pointer size: 131 Bytes
  • Size of remote file: 401 kB
A-表番角色/噬血狂袭/15_Himeragi_Yukina-29.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b6b15072053f101346747650a9d989aecb167db4e19d45ca57cd799fdc52ec2
3
+ size 37877848
A-表番角色/噬血狂袭/15_Kanase_Kanon-30.civitai.info ADDED
@@ -0,0 +1,232 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2326861,
3
+ "modelId": 2056166,
4
+ "name": "v1.0",
5
+ "nsfwLevel": 60,
6
+ "createdAt": "2025-10-19T00:49:39.236Z",
7
+ "updatedAt": "2025-10-19T00:54:25.161Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-19T00:54:25.124Z",
10
+ "trainedWords": [
11
+ "kanase kanon"
12
+ ],
13
+ "trainingStatus": null,
14
+ "trainingDetails": null,
15
+ "baseModel": "SD 1.5",
16
+ "baseModelType": "Standard",
17
+ "earlyAccessEndsAt": null,
18
+ "earlyAccessConfig": null,
19
+ "description": null,
20
+ "uploadType": "Created",
21
+ "usageControl": "Download",
22
+ "air": "urn:air:sd1:lora:civitai:2056166@2326861",
23
+ "stats": {
24
+ "downloadCount": 20,
25
+ "ratingCount": 0,
26
+ "rating": 0,
27
+ "thumbsUpCount": 8
28
+ },
29
+ "model": {
30
+ "name": "Kanase Kanon | Strike the Blood | \u30b9\u30c8\u30e9\u30a4\u30af\u30fb\u30b6\u30fb\u30d6\u30e9\u30c3\u30c9",
31
+ "type": "LORA",
32
+ "nsfw": true,
33
+ "poi": false,
34
+ "description": "Lora-scale: 0.6-0.8\nTrigger Words: kanase kanon, short hair, grey hair, blue eyes\n\n\n\nIf you'd like to support my work, you can do so through Ko-fi!",
35
+ "tags": [
36
+ "anime",
37
+ "character",
38
+ "strike the blood",
39
+ "girls"
40
+ ],
41
+ "allowNoCredit": true,
42
+ "allowCommercialUse": [
43
+ "RentCivit",
44
+ "Rent"
45
+ ],
46
+ "allowDerivatives": true,
47
+ "allowDifferentLicense": true
48
+ },
49
+ "files": [
50
+ {
51
+ "id": 2217057,
52
+ "sizeKB": 36984.359375,
53
+ "name": "15_Kanase_Kanon-30.safetensors",
54
+ "type": "Model",
55
+ "pickleScanResult": "Success",
56
+ "pickleScanMessage": "No Pickle imports",
57
+ "virusScanResult": "Success",
58
+ "virusScanMessage": null,
59
+ "scannedAt": "2025-10-19T00:55:40.771Z",
60
+ "metadata": {
61
+ "format": "SafeTensor",
62
+ "size": null,
63
+ "fp": null
64
+ },
65
+ "hashes": {
66
+ "AutoV1": "2C084AE1",
67
+ "AutoV2": "D08C13F222",
68
+ "SHA256": "D08C13F222075867D6A800CBDF20BDA830CA80F79267A97F8F72A96D9D967938",
69
+ "CRC32": "3D06D0C6",
70
+ "BLAKE3": "C0EAF8404C7185044CCEA7081A7FF15A0B32E711D8E03D6488252E9FC0171C9D",
71
+ "AutoV3": "1D5AD6668574"
72
+ },
73
+ "primary": true,
74
+ "downloadUrl": "https://civitai.com/api/download/models/2326861"
75
+ }
76
+ ],
77
+ "images": [
78
+ {
79
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9dbb27f9-9a66-4d01-a848-be6eb3ffa1f9/original=true/106647396.jpeg",
80
+ "nsfwLevel": 1,
81
+ "width": 512,
82
+ "height": 768,
83
+ "hash": "UjJICV?aI:E1.TRiXAtRSibGj]ofkWxvadRj",
84
+ "type": "image",
85
+ "metadata": {
86
+ "hash": "UjJICV?aI:E1.TRiXAtRSibGj]ofkWxvadRj",
87
+ "size": 483227,
88
+ "width": 512,
89
+ "height": 768,
90
+ "nsfwLevelReason": null
91
+ },
92
+ "minor": false,
93
+ "poi": false,
94
+ "meta": {
95
+ "seed": 973836194769912,
96
+ "Model": "Anyorange1.5",
97
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 973836194769912, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, camera, smile, looking at viewer, bow, shirt, day, white shirt, blue bow, sky, upper body\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Kanase_Kanon-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [973836194769912, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Kanase_Kanon-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, camera, smile, looking at viewer, bow, shirt, day, white shirt, blue bow, sky, upper body\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [184.34286517250624, 15.255886752895323]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
98
+ "steps": 30,
99
+ "width": 512,
100
+ "height": 768,
101
+ "models": [
102
+ "Anyorange1.5.safetensors"
103
+ ],
104
+ "prompt": "masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, camera, smile, looking at viewer, bow, shirt, day, white shirt, blue bow, sky, upper body",
105
+ "denoise": 1,
106
+ "sampler": "Euler a",
107
+ "cfgScale": 5,
108
+ "scheduler": "normal",
109
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
110
+ "additionalResources": [
111
+ {
112
+ "name": "1.5/Kanase_Kanon-30.safetensors",
113
+ "type": "lora",
114
+ "strength": 0.7000000000000002,
115
+ "strengthClip": 1
116
+ }
117
+ ]
118
+ },
119
+ "availability": "Public",
120
+ "hasMeta": true,
121
+ "hasPositivePrompt": true,
122
+ "onSite": false,
123
+ "remixOfId": null
124
+ },
125
+ {
126
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/9ae1037e-a06d-40dd-a379-f57bcc179b96/original=true/106647398.jpeg",
127
+ "nsfwLevel": 1,
128
+ "width": 512,
129
+ "height": 768,
130
+ "hash": "UCJHt3-p00MxxU~VD%9E4n9a0Kxu~qt6ofIo",
131
+ "type": "image",
132
+ "metadata": {
133
+ "hash": "UCJHt3-p00MxxU~VD%9E4n9a0Kxu~qt6ofIo",
134
+ "size": 481713,
135
+ "width": 512,
136
+ "height": 768,
137
+ "nsfwLevelReason": null
138
+ },
139
+ "minor": false,
140
+ "poi": false,
141
+ "meta": {
142
+ "seed": 194947444400135,
143
+ "Model": "Anyorange1.5",
144
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 194947444400135, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, school uniform, open mouth, day, upper body, serafuku\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Kanase_Kanon-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [194947444400135, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Kanase_Kanon-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, school uniform, open mouth, day, upper body, serafuku\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [184.34286517250624, 15.255886752895323]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
145
+ "steps": 30,
146
+ "width": 512,
147
+ "height": 768,
148
+ "models": [
149
+ "Anyorange1.5.safetensors"
150
+ ],
151
+ "prompt": "masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, school uniform, open mouth, day, upper body, serafuku",
152
+ "denoise": 1,
153
+ "sampler": "Euler a",
154
+ "cfgScale": 5,
155
+ "scheduler": "normal",
156
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
157
+ "additionalResources": [
158
+ {
159
+ "name": "1.5/Kanase_Kanon-30.safetensors",
160
+ "type": "lora",
161
+ "strength": 0.7000000000000002,
162
+ "strengthClip": 1
163
+ }
164
+ ]
165
+ },
166
+ "availability": "Public",
167
+ "hasMeta": true,
168
+ "hasPositivePrompt": true,
169
+ "onSite": false,
170
+ "remixOfId": null
171
+ },
172
+ {
173
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/79497da1-ba69-4044-b008-2dffe20c2caa/original=true/106647397.jpeg",
174
+ "nsfwLevel": 1,
175
+ "width": 512,
176
+ "height": 768,
177
+ "hash": "U4JaZf4.00^*00^+L~%K00D*_3o}8w0K_4%N",
178
+ "type": "image",
179
+ "metadata": {
180
+ "hash": "U4JaZf4.00^*00^+L~%K00D*_3o}8w0K_4%N",
181
+ "size": 422466,
182
+ "width": 512,
183
+ "height": 768,
184
+ "nsfwLevelReason": null
185
+ },
186
+ "minor": false,
187
+ "poi": false,
188
+ "meta": {
189
+ "seed": 1124023241271573,
190
+ "Model": "Anyorange1.5",
191
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 1124023241271573, \"steps\": 30, \"cfg\": 5.0, \"sampler_name\": \"euler_ancestral\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"11\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"Anyorange1.5.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 3}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, open mouth, pajamas, braid, shirt, pink shirt, bow, upper body, hair bow\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\", \"clip\": [\"13\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"4\", 2]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"2loras_test_\", \"images\": [\"8\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"11\": {\"inputs\": {\"lora_name\": \"1.5/Kanase_Kanon-30.safetensors\", \"strength_model\": 0.7000000000000002, \"strength_clip\": 1.0000000000000002, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"13\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"11\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}}, \"workflow\": {\"id\": \"268af3b2-7ad6-4b26-8dd4-652a7d7da106\", \"revision\": 0, \"last_node_id\": 13, \"last_link_id\": 24, \"nodes\": [{\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [770, 170], \"size\": [210, 46], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"label\": \"samples\", \"name\": \"samples\", \"type\": \"LATENT\", \"link\": 7}, {\"label\": \"vae\", \"name\": \"vae\", \"type\": \"VAE\", \"link\": 8}], \"outputs\": [{\"label\": \"IMAGE\", \"name\": \"IMAGE\", \"type\": \"IMAGE\", \"slot_index\": 0, \"links\": [9]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"VAEDecode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": []}, {\"id\": 12, \"type\": \"MarkdownNote\", \"pos\": [-776, 432], \"size\": [312, 88], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [], \"properties\": {\"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"### Learn more about this workflow\\n\\n> [LoRA - ComfyUI_examples](https://comfyanonymous.github.io/ComfyUI_examples/lora/) \\u2014 Overview\\n> \\n> [Multiple LoRAs - docs.comfy.org](https://docs.comfy.org/tutorials/basic/multiple-loras) \\u2014 Detailed guide to using multiple LoRAs\"], \"color\": \"#432\", \"bgcolor\": \"#653\"}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [984.2078247070312, 280.6827392578125], \"size\": [210, 270], \"flags\": {}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"label\": \"images\", \"name\": \"images\", \"type\": \"IMAGE\", \"link\": 9}], \"outputs\": [], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"SaveImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"2loras_test_\"]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [420, 170], \"size\": [315, 262], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 19}, {\"label\": \"positive\", \"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"label\": \"negative\", \"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"label\": \"latent_image\", \"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [7]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"KSampler\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [1124023241271573, \"randomize\", 30, 5, \"euler_ancestral\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [-37.28086853027344, 390.07464599609375], \"size\": [425.27801513671875, 88], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 24}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [6]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\\n\"]}, {\"id\": 13, \"type\": \"CLIPSetLastLayer\", \"pos\": [-344.6966247558594, 233.2925262451172], \"size\": [270, 58], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 22}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [23, 24]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.35\", \"Node name for S&R\": \"CLIPSetLastLayer\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [-2]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-776, 288], \"size\": [315, 98], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [17]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [18]}, {\"label\": \"VAE\", \"name\": \"VAE\", \"type\": \"VAE\", \"slot_index\": 2, \"links\": [8]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CheckpointLoaderSimple\", \"models\": [{\"name\": \"dreamshaper_8.safetensors\", \"url\": \"https://civitai.com/api/download/models/128713?type=Model&format=SafeTensor&size=pruned&fp=fp16\", \"directory\": \"checkpoints\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"Anyorange1.5.safetensors\"]}, {\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [52.024810791015625, 641.3909301757812], \"size\": [315, 106], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"label\": \"LATENT\", \"name\": \"LATENT\", \"type\": \"LATENT\", \"slot_index\": 0, \"links\": [2]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"EmptyLatentImage\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [512, 768, 3]}, {\"id\": 11, \"type\": \"LoraLoader\", \"pos\": [-655.799560546875, 37.162010192871094], \"size\": [315, 126], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [{\"label\": \"model\", \"name\": \"model\", \"type\": \"MODEL\", \"link\": 17}, {\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 18}], \"outputs\": [{\"label\": \"MODEL\", \"name\": \"MODEL\", \"type\": \"MODEL\", \"slot_index\": 0, \"links\": [19]}, {\"label\": \"CLIP\", \"name\": \"CLIP\", \"type\": \"CLIP\", \"slot_index\": 1, \"links\": [22]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"LoraLoader\", \"models\": [{\"name\": \"blindbox_v1_mix.safetensors\", \"url\": \"https://civitai.com/api/download/models/32988?type=Model&format=SafeTensor&size=full&fp=fp16\", \"directory\": \"loras\"}], \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"1.5/Kanase_Kanon-30.safetensors\", 0.7000000000000002, 1.0000000000000002]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [-30, 170], \"size\": [422.84503173828125, 88], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"label\": \"clip\", \"name\": \"clip\", \"type\": \"CLIP\", \"link\": 23}], \"outputs\": [{\"label\": \"CONDITIONING\", \"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"slot_index\": 0, \"links\": [4]}], \"properties\": {\"cnr_id\": \"comfy-core\", \"ver\": \"0.3.18\", \"Node name for S&R\": \"CLIPTextEncode\", \"ue_properties\": {\"version\": \"7.0.1\", \"widget_ue_connectable\": {}}}, \"widgets_values\": [\"masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, open mouth, pajamas, braid, shirt, pink shirt, bow, upper body, hair bow\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [7, 3, 0, 8, 0, \"LATENT\"], [8, 4, 2, 8, 1, \"VAE\"], [9, 8, 0, 9, 0, \"IMAGE\"], [17, 4, 0, 11, 0, \"MODEL\"], [18, 4, 1, 11, 1, \"CLIP\"], [19, 11, 0, 3, 0, \"MODEL\"], [22, 11, 1, 13, 0, \"CLIP\"], [23, 13, 0, 6, 0, \"CLIP\"], [24, 13, 0, 7, 0, \"CLIP\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 0.922959981770646, \"offset\": [184.34286517250624, 15.255886752895323]}, \"frontendVersion\": \"1.25.11\", \"node_versions\": {\"comfy-core\": \"v0.3.9\"}, \"ue_links\": [], \"links_added_by_ue\": []}, \"version\": 0.4}}",
192
+ "steps": 30,
193
+ "width": 512,
194
+ "height": 768,
195
+ "models": [
196
+ "Anyorange1.5.safetensors"
197
+ ],
198
+ "prompt": "masterpiece,best quality,amazing quality, kanase kanon, short hair, grey hair, blue eyes, 1girl, solo, open mouth, pajamas, braid, shirt, pink shirt, bow, upper body, hair bow",
199
+ "denoise": 1,
200
+ "sampler": "Euler a",
201
+ "cfgScale": 5,
202
+ "scheduler": "normal",
203
+ "negativePrompt": "bad quality,worst quality,worst detail,sketch,censor, simple background,transparent background\n",
204
+ "additionalResources": [
205
+ {
206
+ "name": "1.5/Kanase_Kanon-30.safetensors",
207
+ "type": "lora",
208
+ "strength": 0.7000000000000002,
209
+ "strengthClip": 1
210
+ }
211
+ ]
212
+ },
213
+ "availability": "Public",
214
+ "hasMeta": true,
215
+ "hasPositivePrompt": true,
216
+ "onSite": false,
217
+ "remixOfId": null
218
+ }
219
+ ],
220
+ "downloadUrl": "https://civitai.com/api/download/models/2326861",
221
+ "creator": {
222
+ "username": "duongve13112002",
223
+ "image": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/f5b736ce-ca77-4637-896d-475236c4196e/width=96/duongve13112002.jpeg"
224
+ },
225
+ "extensions": {
226
+ "sd_civitai_helper": {
227
+ "version": "1.8.13",
228
+ "last_update": 1760877722,
229
+ "skeleton_file": false
230
+ }
231
+ }
232
+ }
A-表番角色/噬血狂袭/15_Kanase_Kanon-30.preview.png ADDED

Git LFS Details

  • SHA256: d3ee70d183d9540a394d4104c554ff786aa36d2b397f9a9174e9604c7ca00b36
  • Pointer size: 131 Bytes
  • Size of remote file: 483 kB
A-表番角色/噬血狂袭/15_Kanase_Kanon-30.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d08c13f222075867d6a800cbdf20bda830ca80f79267a97f8f72a96d9d967938
3
+ size 37871984
A-表番角色/噬血狂袭/KirasakaSayakaStrike_v10.preview.png CHANGED

Git LFS Details

  • SHA256: b8f584d93d72acce31d34daefcd26e07e39e5a22081e1c91accaabe81d328157
  • Pointer size: 131 Bytes
  • Size of remote file: 483 kB
A-表番角色/噬血狂袭/asagi_aiba.preview.png CHANGED

Git LFS Details

  • SHA256: a5d69d10861bc607c6dbd1dd506527a72f683fe4e6ff6eb83cc18fe7d8eb3800
  • Pointer size: 131 Bytes
  • Size of remote file: 153 kB
A-表番角色/噬血狂袭/kanaseKanonStrikeTheBlood_v10.preview.png CHANGED

Git LFS Details

  • SHA256: c08186d72d7eb497c017cef89671b1a1bacb76f3c26d8af79f1d01ee3336c043
  • Pointer size: 131 Bytes
  • Size of remote file: 443 kB
A-表番角色/噬血狂袭/la_folia.preview.png CHANGED

Git LFS Details

  • SHA256: 9bebb48c71e270d352a1d2c6fa831749e96553106505f7b0ffe780ecd4532dbe
  • Pointer size: 131 Bytes
  • Size of remote file: 150 kB
A-表番角色/噬血狂袭/natsuki_minamiya.preview.png CHANGED

Git LFS Details

  • SHA256: febb402ce5b505a2065bed9c81d69e33d8b92dd241c492de33026ca1a4c6a515
  • Pointer size: 131 Bytes
  • Size of remote file: 162 kB
A-表番角色/噬血狂袭/natsuki_minamiya_no_outfit.preview.png CHANGED

Git LFS Details

  • SHA256: 90137e07364218f859e8213bfd2079ba393bd00f660c647aee49891abb44a2a9
  • Pointer size: 131 Bytes
  • Size of remote file: 168 kB
A-表番角色/噬血狂袭/yukina_himeragi.preview.png CHANGED

Git LFS Details

  • SHA256: 802f29c09e8a0fcce61ada38eacf3094914471de0931506f9040638f85beecf1
  • Pointer size: 131 Bytes
  • Size of remote file: 137 kB
A-表番角色/噬血狂袭/yukina_himeragi_no_outfit.preview.png CHANGED

Git LFS Details

  • SHA256: 15a1a2f9d1085b59182948aee0161b198a510c502099500af352da64553b89db
  • Pointer size: 131 Bytes
  • Size of remote file: 158 kB
A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.civitai.info ADDED
@@ -0,0 +1,291 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2312694,
3
+ "modelId": 2043366,
4
+ "name": "V1",
5
+ "nsfwLevel": 2,
6
+ "createdAt": "2025-10-14T10:25:15.780Z",
7
+ "updatedAt": "2025-10-14T11:41:14.200Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-14T11:41:14.175Z",
10
+ "trainedWords": [],
11
+ "trainingStatus": "Approved",
12
+ "trainingDetails": {
13
+ "type": "Character",
14
+ "params": {
15
+ "engine": "kohya",
16
+ "unetLR": 0.0005,
17
+ "clipSkip": 2,
18
+ "loraType": "lora",
19
+ "keepTokens": 0,
20
+ "networkDim": 16,
21
+ "numRepeats": 30,
22
+ "resolution": 512,
23
+ "lrScheduler": "cosine_with_restarts",
24
+ "minSnrGamma": 5,
25
+ "noiseOffset": 0.1,
26
+ "targetSteps": 500,
27
+ "enableBucket": true,
28
+ "networkAlpha": 8,
29
+ "optimizerType": "AdamW8Bit",
30
+ "textEncoderLR": 0.0001,
31
+ "maxTrainEpochs": 10,
32
+ "shuffleCaption": false,
33
+ "trainBatchSize": 6,
34
+ "flipAugmentation": false,
35
+ "lrSchedulerNumCycles": 3
36
+ },
37
+ "staging": false,
38
+ "baseModel": "anime",
39
+ "mediaType": "image",
40
+ "highPriority": false,
41
+ "baseModelType": "sd15",
42
+ "samplePrompts": [
43
+ "",
44
+ "",
45
+ ""
46
+ ],
47
+ "negativePrompt": "bad quality, low quality, worst quality, jpeg artifacts, blurry, pixelated, out of focus, watermark, text, signature"
48
+ },
49
+ "baseModel": "SD 1.5",
50
+ "baseModelType": "Standard",
51
+ "earlyAccessEndsAt": null,
52
+ "earlyAccessConfig": null,
53
+ "description": null,
54
+ "uploadType": "Trained",
55
+ "usageControl": "Download",
56
+ "air": "urn:air:sd1:lora:civitai:2043366@2312694",
57
+ "stats": {
58
+ "downloadCount": 39,
59
+ "ratingCount": 0,
60
+ "rating": 0,
61
+ "thumbsUpCount": 6
62
+ },
63
+ "model": {
64
+ "name": "Orphia - Seirei Gensouki",
65
+ "type": "LORA",
66
+ "nsfw": false,
67
+ "poi": false,
68
+ "description": "Orphia - Seirei Gensouki",
69
+ "tags": [
70
+ "character"
71
+ ],
72
+ "allowNoCredit": true,
73
+ "allowCommercialUse": [],
74
+ "allowDerivatives": true,
75
+ "allowDifferentLicense": true
76
+ },
77
+ "files": [
78
+ {
79
+ "id": 2203276,
80
+ "sizeKB": 18546.8515625,
81
+ "name": "Orphia_-_Seirei_Gensouki.safetensors",
82
+ "type": "Model",
83
+ "pickleScanResult": "Success",
84
+ "pickleScanMessage": "No Pickle imports",
85
+ "virusScanResult": "Success",
86
+ "virusScanMessage": null,
87
+ "scannedAt": "2025-10-14T11:00:46.938Z",
88
+ "metadata": {
89
+ "format": "SafeTensor"
90
+ },
91
+ "hashes": {
92
+ "AutoV1": "DE2F2560",
93
+ "AutoV2": "B5C28C50F6",
94
+ "SHA256": "B5C28C50F6B905D8B902AF1D25282625EFC46480E1586D72200AF20A92319485",
95
+ "CRC32": "665305DA",
96
+ "BLAKE3": "B00642E241DD6B9A1B0E1506B84BBD6630D7B95EE5F5D267C23EF1F6B3853FE7",
97
+ "AutoV3": "7220F08A5F6D"
98
+ },
99
+ "primary": true,
100
+ "downloadUrl": "https://civitai.com/api/download/models/2312694"
101
+ }
102
+ ],
103
+ "images": [
104
+ {
105
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/b6db5d43-2bbc-4f61-8728-ca837ae9f5ce/original=true/106015975.jpeg",
106
+ "nsfwLevel": 2,
107
+ "width": 2048,
108
+ "height": 3072,
109
+ "hash": "UHH{ZUo}HbxYHXWC00V?}]flOuaeC4bI~Bs.",
110
+ "type": "image",
111
+ "metadata": {
112
+ "hash": "UHH{ZUo}HbxYHXWC00V?}]flOuaeC4bI~Bs.",
113
+ "size": 6358251,
114
+ "width": 2048,
115
+ "height": 3072
116
+ },
117
+ "minor": false,
118
+ "poi": false,
119
+ "meta": {
120
+ "seed": 427861574490801,
121
+ "vaes": [
122
+ "CleanVAE.safetensors"
123
+ ],
124
+ "Model": "falkonsAnimeAndHentai_v13",
125
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 427861574490801, \"steps\": 10, \"cfg\": 2.0, \"sampler_name\": \"lcm\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"160\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"falkonsAnimeAndHentai_v13.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 1}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \\n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \\n(bedroom, on bed, sitting:1.3), (looking at viewer, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"164\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"157\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"33\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"160\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}, \"157\": {\"inputs\": {\"upscale_model\": [\"158\", 0], \"image\": [\"8\", 0]}, \"class_type\": \"ImageUpscaleWithModel\", \"_meta\": {\"title\": \"Upscale Image (using Model)\"}}, \"158\": {\"inputs\": {\"model_name\": \"4xUltrasharp_4xUltrasharpV10.pt\"}, \"class_type\": \"UpscaleModelLoader\", \"_meta\": {\"title\": \"Load Upscale Model\"}}, \"159\": {\"inputs\": {\"lora_name\": \"age_slider_v20.safetensors\", \"strength_model\": 0.0, \"strength_clip\": 1.0, \"model\": [\"161\", 0], \"clip\": [\"161\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"160\": {\"inputs\": {\"lora_name\": \"PAseer-SD15-LCM Quick.safetensors\", \"strength_model\": 1.0, \"strength_clip\": 1.0, \"model\": [\"159\", 0], \"clip\": [\"159\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"161\": {\"inputs\": {\"lora_name\": \"Orphia_-_Seirei_Gensouki.safetensors\", \"strength_model\": 0.6, \"strength_clip\": 1.0, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"163\": {\"inputs\": {\"lora_name\": \"Mesa_Milkano_-_Swordmasters_Youngest_Son.safetensors\", \"Base Model\": \"SD 1.5\", \"output\": \"URL: https://civitai.com/models/2037308\\nTriggers: mesa milkano\\nBase Model: SD 1.5\\n\\nExamples:\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6439eed5-1859-496f-8057-be09ba188650/original=true/105649114.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: breastplate, greaves, petals, looking_back, green_eyes, blunt_bangs, pink_hair, brown_belt, day, 1girl, bangs, night_sky, pants, arms_at_sides, star_\\\\(sky\\\\), gauntlets, sunset, sidelocks, blue_sky, night, belt, wariza, pov, black_pants, blush, vambraces, ponytail, katou_asuka\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e080074c-6edd-41d9-aa00-aad6068ef827/original=true/105649119.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: hands_on_own_knees, pov, black_gloves, bush, ponytail, open_mouth, mountain, brown_belt, floating_hair, solo, pink_hair, sidelocks, forest, looking_to_the_side, flower_field, crossed_legs, armored_boots, closed_mouth, petals, day, blunt_bangs, armor, field, long_hair, bangs, looking_back, night_sky, nature\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/15419230-62c5-4ee7-9465-4b102e8a0071/original=true/105649111.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: nature, gloves, outdoors, ponytail, petals, looking_back, armor, , mountain, looking_at_viewer, forest, brown_belt, bush, hands_on_own_knees, alternate_hairstyle, tree, green_eyes, star_\\\\(sky\\\\), sky, sitting, grass, sunset, looking_to_the_side, bangs, cowboy_shot, blunt_bangs, smile, from_behind\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\"}, \"class_type\": \"LoraInfo\", \"_meta\": {\"title\": \"Lora Info\"}}, \"164\": {\"inputs\": {\"vae_name\": \"CleanVAE.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}}, \"workflow\": {\"last_node_id\": 164, \"last_link_id\": 794, \"nodes\": [{\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [629.10693359375, 235.40773010253906], \"size\": [315, 106], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [512, 768, 1]}, {\"id\": 157, \"type\": \"ImageUpscaleWithModel\", \"pos\": [1911.0567626953125, -203.64340209960938], \"size\": [340.20001220703125, 46], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"upscale_model\", \"type\": \"UPSCALE_MODEL\", \"link\": 766}, {\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 767}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [768], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"ImageUpscaleWithModel\"}, \"widgets_values\": []}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [2322.21484375, -169.14207458496094], \"size\": [364.16668701171875, 404.3333740234375], \"flags\": {\"collapsed\": true}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 768}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1580.349365234375, -17.497373580932617], \"size\": [210, 46], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 262}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 794}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [767], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 158, \"type\": \"UpscaleModelLoader\", \"pos\": [1537.1036376953125, -273.8653259277344], \"size\": [315, 58], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"UPSCALE_MODEL\", \"type\": \"UPSCALE_MODEL\", \"links\": [766], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UpscaleModelLoader\"}, \"widgets_values\": [\"4xUltrasharp_4xUltrasharpV10.pt\"]}, {\"id\": 33, \"type\": \"CLIPSetLastLayer\", \"pos\": [66.87496948242188, 47.42504119873047], \"size\": [315, 58], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 775}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [792, 793], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPSetLastLayer\"}, \"widgets_values\": [-2]}, {\"id\": 164, \"type\": \"VAELoader\", \"pos\": [1144.0006103515625, 76.65125274658203], \"size\": [315, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [794], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"CleanVAE.safetensors\"]}, {\"id\": 160, \"type\": \"LoraLoader\", \"pos\": [58.979976654052734, -166.80343627929688], \"size\": [315, 126], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 772}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 787}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [791], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [775], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"PAseer-SD15-LCM Quick.safetensors\", 1, 1]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [1130.9888916015625, -342.4229431152344], \"size\": [315, 262], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 791}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [262], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [427861574490801, \"randomize\", 10, 2, \"lcm\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [528.8944091796875, -10.626336097717285], \"size\": [425.27801513671875, 180.6060791015625], \"flags\": {\"collapsed\": false}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 793}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)\"]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-317.33526611328125, -555.9864501953125], \"size\": [315, 98], \"flags\": {\"collapsed\": false}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [780], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [781], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"falkonsAnimeAndHentai_v13.safetensors\"]}, {\"id\": 163, \"type\": \"LoraInfo\", \"pos\": [534.1552734375, -668.4848022460938], \"size\": [346.52117919921875, 293.06060791015625], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"lora_name\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"trigger_words\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"example_prompt\", \"type\": \"STRING\", \"links\": null}], \"properties\": {\"Node name for S&R\": \"LoraInfo\"}, \"widgets_values\": [\"Mesa_Milkano_-_Swordmasters_Youngest_Son.safetensors\", \"SD 1.5\", \"URL: https://civitai.com/models/2037308\\nTriggers: mesa milkano\\nBase Model: SD 1.5\\n\\nExamples:\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6439eed5-1859-496f-8057-be09ba188650/original=true/105649114.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: breastplate, greaves, petals, looking_back, green_eyes, blunt_bangs, pink_hair, brown_belt, day, 1girl, bangs, night_sky, pants, arms_at_sides, star_\\\\(sky\\\\), gauntlets, sunset, sidelocks, blue_sky, night, belt, wariza, pov, black_pants, blush, vambraces, ponytail, katou_asuka\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e080074c-6edd-41d9-aa00-aad6068ef827/original=true/105649119.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: hands_on_own_knees, pov, black_gloves, bush, ponytail, open_mouth, mountain, brown_belt, floating_hair, solo, pink_hair, sidelocks, forest, looking_to_the_side, flower_field, crossed_legs, armored_boots, closed_mouth, petals, day, blunt_bangs, armor, field, long_hair, bangs, looking_back, night_sky, nature\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/15419230-62c5-4ee7-9465-4b102e8a0071/original=true/105649111.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: nature, gloves, outdoors, ponytail, petals, looking_back, armor, , mountain, looking_at_viewer, forest, brown_belt, bush, hands_on_own_knees, alternate_hairstyle, tree, green_eyes, star_\\\\(sky\\\\), sky, sitting, grass, sunset, looking_to_the_side, bangs, cowboy_shot, blunt_bangs, smile, from_behind\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\"]}, {\"id\": 161, \"type\": \"LoraLoader\", \"pos\": [60.91288375854492, -547.70703125], \"size\": [315, 126], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 780}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 781}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [786], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [790], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"Orphia_-_Seirei_Gensouki.safetensors\", 0.6, 1]}, {\"id\": 159, \"type\": \"LoraLoader\", \"pos\": [55.07668685913086, -364.66888427734375], \"size\": [315, 126], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 786}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 790}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [772], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [787], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"age_slider_v20.safetensors\", 0, 1]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [531.701904296875, -299.3091735839844], \"size\": [423.61407470703125, 228.15113830566406], \"flags\": {\"collapsed\": false}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 792}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \\n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \\n(bedroom, on bed, sitting:1.3), (looking at viewer, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [262, 3, 0, 8, 0, \"LATENT\"], [766, 158, 0, 157, 0, \"UPSCALE_MODEL\"], [767, 8, 0, 157, 1, \"IMAGE\"], [768, 157, 0, 9, 0, \"IMAGE\"], [772, 159, 0, 160, 0, \"MODEL\"], [775, 160, 1, 33, 0, \"CLIP\"], [780, 4, 0, 161, 0, \"MODEL\"], [781, 4, 1, 161, 1, \"CLIP\"], [786, 161, 0, 159, 0, \"MODEL\"], [787, 159, 1, 160, 1, \"CLIP\"], [790, 161, 1, 159, 1, \"CLIP\"], [791, 160, 0, 3, 0, \"MODEL\"], [792, 33, 0, 6, 0, \"CLIP\"], [793, 33, 0, 7, 0, \"CLIP\"], [794, 164, 0, 8, 1, \"VAE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1918176537727436, \"offset\": [293.1093270451885, 595.4901553635286]}, \"node_versions\": {\"comfy-core\": \"0.3.10\", \"lora-info\": \"5db14b8f37e65585d6114416feb8f7b110541548\"}}, \"version\": 0.4}}",
126
+ "steps": 10,
127
+ "width": 512,
128
+ "height": 768,
129
+ "models": [
130
+ "falkonsAnimeAndHentai_v13.safetensors"
131
+ ],
132
+ "prompt": "(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \n(bedroom, on bed, sitting:1.3), (looking at viewer, eye level shot:1.28), \n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)",
133
+ "denoise": 1,
134
+ "sampler": "LCM",
135
+ "cfgScale": 2,
136
+ "scheduler": "normal",
137
+ "upscalers": [
138
+ "4xUltrasharp_4xUltrasharpV10.pt"
139
+ ],
140
+ "negativePrompt": "embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)",
141
+ "additionalResources": [
142
+ {
143
+ "name": "PAseer-SD15-LCM Quick.safetensors",
144
+ "type": "lora",
145
+ "strength": 1,
146
+ "strengthClip": 1
147
+ },
148
+ {
149
+ "name": "Orphia_-_Seirei_Gensouki.safetensors",
150
+ "type": "lora",
151
+ "strength": 0.6,
152
+ "strengthClip": 1
153
+ }
154
+ ]
155
+ },
156
+ "availability": "Public",
157
+ "hasMeta": true,
158
+ "hasPositivePrompt": true,
159
+ "onSite": false,
160
+ "remixOfId": null
161
+ },
162
+ {
163
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/d65af866-3d84-4b9d-88d0-47d19b62b7be/original=true/106016113.jpeg",
164
+ "nsfwLevel": 2,
165
+ "width": 2048,
166
+ "height": 3072,
167
+ "hash": "USE3kvR-02oL_KR+NLs:?DR+OFoL?ZR-NGjs",
168
+ "type": "image",
169
+ "metadata": {
170
+ "hash": "USE3kvR-02oL_KR+NLs:?DR+OFoL?ZR-NGjs",
171
+ "size": 7887080,
172
+ "width": 2048,
173
+ "height": 3072
174
+ },
175
+ "minor": false,
176
+ "poi": false,
177
+ "meta": {
178
+ "seed": 586803840459351,
179
+ "vaes": [
180
+ "CleanVAE.safetensors"
181
+ ],
182
+ "Model": "anyloracleanlinearmix_v10",
183
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 586803840459351, \"steps\": 10, \"cfg\": 2.0, \"sampler_name\": \"lcm\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"160\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"anyloracleanlinearmix_v10.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 1}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \\n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \\n(outdoor, at night, standing, full moon, stars:1.3), (looking at viewer, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"164\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"157\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"33\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"160\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}, \"157\": {\"inputs\": {\"upscale_model\": [\"158\", 0], \"image\": [\"8\", 0]}, \"class_type\": \"ImageUpscaleWithModel\", \"_meta\": {\"title\": \"Upscale Image (using Model)\"}}, \"158\": {\"inputs\": {\"model_name\": \"4xUltrasharp_4xUltrasharpV10.pt\"}, \"class_type\": \"UpscaleModelLoader\", \"_meta\": {\"title\": \"Load Upscale Model\"}}, \"159\": {\"inputs\": {\"lora_name\": \"age_slider_v20.safetensors\", \"strength_model\": 0.0, \"strength_clip\": 1.0, \"model\": [\"161\", 0], \"clip\": [\"161\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"160\": {\"inputs\": {\"lora_name\": \"PAseer-SD15-LCM Quick.safetensors\", \"strength_model\": 1.0, \"strength_clip\": 1.0, \"model\": [\"159\", 0], \"clip\": [\"159\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"161\": {\"inputs\": {\"lora_name\": \"Orphia_-_Seirei_Gensouki.safetensors\", \"strength_model\": 0.6, \"strength_clip\": 1.0, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"163\": {\"inputs\": {\"lora_name\": \"Mesa_Milkano_-_Swordmasters_Youngest_Son.safetensors\", \"Base Model\": \"SD 1.5\", \"output\": \"URL: https://civitai.com/models/2037308\\nTriggers: mesa milkano\\nBase Model: SD 1.5\\n\\nExamples:\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6439eed5-1859-496f-8057-be09ba188650/original=true/105649114.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: breastplate, greaves, petals, looking_back, green_eyes, blunt_bangs, pink_hair, brown_belt, day, 1girl, bangs, night_sky, pants, arms_at_sides, star_\\\\(sky\\\\), gauntlets, sunset, sidelocks, blue_sky, night, belt, wariza, pov, black_pants, blush, vambraces, ponytail, katou_asuka\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e080074c-6edd-41d9-aa00-aad6068ef827/original=true/105649119.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: hands_on_own_knees, pov, black_gloves, bush, ponytail, open_mouth, mountain, brown_belt, floating_hair, solo, pink_hair, sidelocks, forest, looking_to_the_side, flower_field, crossed_legs, armored_boots, closed_mouth, petals, day, blunt_bangs, armor, field, long_hair, bangs, looking_back, night_sky, nature\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/15419230-62c5-4ee7-9465-4b102e8a0071/original=true/105649111.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: nature, gloves, outdoors, ponytail, petals, looking_back, armor, , mountain, looking_at_viewer, forest, brown_belt, bush, hands_on_own_knees, alternate_hairstyle, tree, green_eyes, star_\\\\(sky\\\\), sky, sitting, grass, sunset, looking_to_the_side, bangs, cowboy_shot, blunt_bangs, smile, from_behind\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\"}, \"class_type\": \"LoraInfo\", \"_meta\": {\"title\": \"Lora Info\"}}, \"164\": {\"inputs\": {\"vae_name\": \"CleanVAE.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}}, \"workflow\": {\"last_node_id\": 164, \"last_link_id\": 794, \"nodes\": [{\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [629.10693359375, 235.40773010253906], \"size\": [315, 106], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [512, 768, 1]}, {\"id\": 157, \"type\": \"ImageUpscaleWithModel\", \"pos\": [1911.0567626953125, -203.64340209960938], \"size\": [340.20001220703125, 46], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"upscale_model\", \"type\": \"UPSCALE_MODEL\", \"link\": 766}, {\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 767}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [768], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"ImageUpscaleWithModel\"}, \"widgets_values\": []}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [2322.21484375, -169.14207458496094], \"size\": [364.16668701171875, 404.3333740234375], \"flags\": {\"collapsed\": true}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 768}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1580.349365234375, -17.497373580932617], \"size\": [210, 46], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 262}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 794}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [767], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 158, \"type\": \"UpscaleModelLoader\", \"pos\": [1537.1036376953125, -273.8653259277344], \"size\": [315, 58], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"UPSCALE_MODEL\", \"type\": \"UPSCALE_MODEL\", \"links\": [766], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UpscaleModelLoader\"}, \"widgets_values\": [\"4xUltrasharp_4xUltrasharpV10.pt\"]}, {\"id\": 33, \"type\": \"CLIPSetLastLayer\", \"pos\": [66.87496948242188, 47.42504119873047], \"size\": [315, 58], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 775}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [792, 793], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPSetLastLayer\"}, \"widgets_values\": [-2]}, {\"id\": 164, \"type\": \"VAELoader\", \"pos\": [1144.0006103515625, 76.65125274658203], \"size\": [315, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [794], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"CleanVAE.safetensors\"]}, {\"id\": 160, \"type\": \"LoraLoader\", \"pos\": [58.979976654052734, -166.80343627929688], \"size\": [315, 126], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 772}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 787}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [791], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [775], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"PAseer-SD15-LCM Quick.safetensors\", 1, 1]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [1130.9888916015625, -342.4229431152344], \"size\": [315, 262], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 791}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [262], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [586803840459351, \"randomize\", 10, 2, \"lcm\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [528.8944091796875, -10.626336097717285], \"size\": [425.27801513671875, 180.6060791015625], \"flags\": {\"collapsed\": false}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 793}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)\"]}, {\"id\": 163, \"type\": \"LoraInfo\", \"pos\": [534.1552734375, -668.4848022460938], \"size\": [346.52117919921875, 293.06060791015625], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"lora_name\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"trigger_words\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"example_prompt\", \"type\": \"STRING\", \"links\": null}], \"properties\": {\"Node name for S&R\": \"LoraInfo\"}, \"widgets_values\": [\"Mesa_Milkano_-_Swordmasters_Youngest_Son.safetensors\", \"SD 1.5\", \"URL: https://civitai.com/models/2037308\\nTriggers: mesa milkano\\nBase Model: SD 1.5\\n\\nExamples:\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6439eed5-1859-496f-8057-be09ba188650/original=true/105649114.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: breastplate, greaves, petals, looking_back, green_eyes, blunt_bangs, pink_hair, brown_belt, day, 1girl, bangs, night_sky, pants, arms_at_sides, star_\\\\(sky\\\\), gauntlets, sunset, sidelocks, blue_sky, night, belt, wariza, pov, black_pants, blush, vambraces, ponytail, katou_asuka\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e080074c-6edd-41d9-aa00-aad6068ef827/original=true/105649119.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: hands_on_own_knees, pov, black_gloves, bush, ponytail, open_mouth, mountain, brown_belt, floating_hair, solo, pink_hair, sidelocks, forest, looking_to_the_side, flower_field, crossed_legs, armored_boots, closed_mouth, petals, day, blunt_bangs, armor, field, long_hair, bangs, looking_back, night_sky, nature\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/15419230-62c5-4ee7-9465-4b102e8a0071/original=true/105649111.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: nature, gloves, outdoors, ponytail, petals, looking_back, armor, , mountain, looking_at_viewer, forest, brown_belt, bush, hands_on_own_knees, alternate_hairstyle, tree, green_eyes, star_\\\\(sky\\\\), sky, sitting, grass, sunset, looking_to_the_side, bangs, cowboy_shot, blunt_bangs, smile, from_behind\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\"]}, {\"id\": 161, \"type\": \"LoraLoader\", \"pos\": [60.91288375854492, -547.70703125], \"size\": [315, 126], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 780}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 781}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [786], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [790], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"Orphia_-_Seirei_Gensouki.safetensors\", 0.6, 1]}, {\"id\": 159, \"type\": \"LoraLoader\", \"pos\": [55.07668685913086, -364.66888427734375], \"size\": [315, 126], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 786}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 790}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [772], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [787], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"age_slider_v20.safetensors\", 0, 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-317.33526611328125, -555.9864501953125], \"size\": [315, 98], \"flags\": {\"collapsed\": false}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [780], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [781], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"anyloracleanlinearmix_v10.safetensors\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [531.701904296875, -299.3091735839844], \"size\": [423.61407470703125, 228.15113830566406], \"flags\": {\"collapsed\": false}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 792}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \\n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \\n(outdoor, at night, standing, full moon, stars:1.3), (looking at viewer, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [262, 3, 0, 8, 0, \"LATENT\"], [766, 158, 0, 157, 0, \"UPSCALE_MODEL\"], [767, 8, 0, 157, 1, \"IMAGE\"], [768, 157, 0, 9, 0, \"IMAGE\"], [772, 159, 0, 160, 0, \"MODEL\"], [775, 160, 1, 33, 0, \"CLIP\"], [780, 4, 0, 161, 0, \"MODEL\"], [781, 4, 1, 161, 1, \"CLIP\"], [786, 161, 0, 159, 0, \"MODEL\"], [787, 159, 1, 160, 1, \"CLIP\"], [790, 161, 1, 159, 1, \"CLIP\"], [791, 160, 0, 3, 0, \"MODEL\"], [792, 33, 0, 6, 0, \"CLIP\"], [793, 33, 0, 7, 0, \"CLIP\"], [794, 164, 0, 8, 1, \"VAE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1918176537727436, \"offset\": [293.1093270451885, 595.4901553635286]}, \"node_versions\": {\"comfy-core\": \"0.3.10\", \"lora-info\": \"5db14b8f37e65585d6114416feb8f7b110541548\"}}, \"version\": 0.4}}",
184
+ "steps": 10,
185
+ "width": 512,
186
+ "height": 768,
187
+ "models": [
188
+ "anyloracleanlinearmix_v10.safetensors"
189
+ ],
190
+ "prompt": "(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \n(outdoor, at night, standing, full moon, stars:1.3), (looking at viewer, eye level shot:1.28), \n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)",
191
+ "denoise": 1,
192
+ "sampler": "LCM",
193
+ "cfgScale": 2,
194
+ "scheduler": "normal",
195
+ "upscalers": [
196
+ "4xUltrasharp_4xUltrasharpV10.pt"
197
+ ],
198
+ "negativePrompt": "embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)",
199
+ "additionalResources": [
200
+ {
201
+ "name": "PAseer-SD15-LCM Quick.safetensors",
202
+ "type": "lora",
203
+ "strength": 1,
204
+ "strengthClip": 1
205
+ },
206
+ {
207
+ "name": "Orphia_-_Seirei_Gensouki.safetensors",
208
+ "type": "lora",
209
+ "strength": 0.6,
210
+ "strengthClip": 1
211
+ }
212
+ ]
213
+ },
214
+ "availability": "Public",
215
+ "hasMeta": true,
216
+ "hasPositivePrompt": true,
217
+ "onSite": false,
218
+ "remixOfId": null
219
+ },
220
+ {
221
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e9c47706-7ab9-4bf7-8526-cbabc5af710b/original=true/106016383.jpeg",
222
+ "nsfwLevel": 2,
223
+ "width": 2048,
224
+ "height": 3072,
225
+ "hash": "U6G[ly%L4T#R?Zoy04M|00WC*JjF0QWX#RoJ",
226
+ "type": "image",
227
+ "metadata": {
228
+ "hash": "U6G[ly%L4T#R?Zoy04M|00WC*JjF0QWX#RoJ",
229
+ "size": 6446394,
230
+ "width": 2048,
231
+ "height": 3072
232
+ },
233
+ "minor": false,
234
+ "poi": false,
235
+ "meta": {
236
+ "seed": 1092263697046330,
237
+ "vaes": [
238
+ "CleanVAE.safetensors"
239
+ ],
240
+ "Model": "anyloracleanlinearmix_v10",
241
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 1092263697046330, \"steps\": 10, \"cfg\": 2.0, \"sampler_name\": \"lcm\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"160\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"anyloracleanlinearmix_v10.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 1}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \\n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \\n(indoor, dining room, sitting on chair, holding cup of tea:1.3), (looking at viewer, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"164\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"157\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"33\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"160\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}, \"157\": {\"inputs\": {\"upscale_model\": [\"158\", 0], \"image\": [\"8\", 0]}, \"class_type\": \"ImageUpscaleWithModel\", \"_meta\": {\"title\": \"Upscale Image (using Model)\"}}, \"158\": {\"inputs\": {\"model_name\": \"4xUltrasharp_4xUltrasharpV10.pt\"}, \"class_type\": \"UpscaleModelLoader\", \"_meta\": {\"title\": \"Load Upscale Model\"}}, \"159\": {\"inputs\": {\"lora_name\": \"age_slider_v20.safetensors\", \"strength_model\": 0.0, \"strength_clip\": 1.0, \"model\": [\"161\", 0], \"clip\": [\"161\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"160\": {\"inputs\": {\"lora_name\": \"PAseer-SD15-LCM Quick.safetensors\", \"strength_model\": 1.0, \"strength_clip\": 1.0, \"model\": [\"159\", 0], \"clip\": [\"159\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"161\": {\"inputs\": {\"lora_name\": \"Orphia_-_Seirei_Gensouki.safetensors\", \"strength_model\": 0.6, \"strength_clip\": 1.0, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"163\": {\"inputs\": {\"lora_name\": \"Mesa_Milkano_-_Swordmasters_Youngest_Son.safetensors\", \"Base Model\": \"SD 1.5\", \"output\": \"URL: https://civitai.com/models/2037308\\nTriggers: mesa milkano\\nBase Model: SD 1.5\\n\\nExamples:\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6439eed5-1859-496f-8057-be09ba188650/original=true/105649114.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: breastplate, greaves, petals, looking_back, green_eyes, blunt_bangs, pink_hair, brown_belt, day, 1girl, bangs, night_sky, pants, arms_at_sides, star_\\\\(sky\\\\), gauntlets, sunset, sidelocks, blue_sky, night, belt, wariza, pov, black_pants, blush, vambraces, ponytail, katou_asuka\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e080074c-6edd-41d9-aa00-aad6068ef827/original=true/105649119.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: hands_on_own_knees, pov, black_gloves, bush, ponytail, open_mouth, mountain, brown_belt, floating_hair, solo, pink_hair, sidelocks, forest, looking_to_the_side, flower_field, crossed_legs, armored_boots, closed_mouth, petals, day, blunt_bangs, armor, field, long_hair, bangs, looking_back, night_sky, nature\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/15419230-62c5-4ee7-9465-4b102e8a0071/original=true/105649111.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: nature, gloves, outdoors, ponytail, petals, looking_back, armor, , mountain, looking_at_viewer, forest, brown_belt, bush, hands_on_own_knees, alternate_hairstyle, tree, green_eyes, star_\\\\(sky\\\\), sky, sitting, grass, sunset, looking_to_the_side, bangs, cowboy_shot, blunt_bangs, smile, from_behind\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\"}, \"class_type\": \"LoraInfo\", \"_meta\": {\"title\": \"Lora Info\"}}, \"164\": {\"inputs\": {\"vae_name\": \"CleanVAE.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}}, \"workflow\": {\"last_node_id\": 164, \"last_link_id\": 794, \"nodes\": [{\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [629.10693359375, 235.40773010253906], \"size\": [315, 106], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [512, 768, 1]}, {\"id\": 157, \"type\": \"ImageUpscaleWithModel\", \"pos\": [1911.0567626953125, -203.64340209960938], \"size\": [340.20001220703125, 46], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"upscale_model\", \"type\": \"UPSCALE_MODEL\", \"link\": 766}, {\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 767}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [768], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"ImageUpscaleWithModel\"}, \"widgets_values\": []}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [2322.21484375, -169.14207458496094], \"size\": [364.16668701171875, 404.3333740234375], \"flags\": {\"collapsed\": true}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 768}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1580.349365234375, -17.497373580932617], \"size\": [210, 46], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 262}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 794}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [767], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 158, \"type\": \"UpscaleModelLoader\", \"pos\": [1537.1036376953125, -273.8653259277344], \"size\": [315, 58], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"UPSCALE_MODEL\", \"type\": \"UPSCALE_MODEL\", \"links\": [766], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UpscaleModelLoader\"}, \"widgets_values\": [\"4xUltrasharp_4xUltrasharpV10.pt\"]}, {\"id\": 33, \"type\": \"CLIPSetLastLayer\", \"pos\": [66.87496948242188, 47.42504119873047], \"size\": [315, 58], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 775}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [792, 793], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPSetLastLayer\"}, \"widgets_values\": [-2]}, {\"id\": 164, \"type\": \"VAELoader\", \"pos\": [1144.0006103515625, 76.65125274658203], \"size\": [315, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [794], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"CleanVAE.safetensors\"]}, {\"id\": 160, \"type\": \"LoraLoader\", \"pos\": [58.979976654052734, -166.80343627929688], \"size\": [315, 126], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 772}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 787}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [791], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [775], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"PAseer-SD15-LCM Quick.safetensors\", 1, 1]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [1130.9888916015625, -342.4229431152344], \"size\": [315, 262], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 791}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [262], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [1092263697046330, \"randomize\", 10, 2, \"lcm\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [528.8944091796875, -10.626336097717285], \"size\": [425.27801513671875, 180.6060791015625], \"flags\": {\"collapsed\": false}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 793}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)\"]}, {\"id\": 163, \"type\": \"LoraInfo\", \"pos\": [534.1552734375, -668.4848022460938], \"size\": [346.52117919921875, 293.06060791015625], \"flags\": {}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"lora_name\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"trigger_words\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"example_prompt\", \"type\": \"STRING\", \"links\": null}], \"properties\": {\"Node name for S&R\": \"LoraInfo\"}, \"widgets_values\": [\"Mesa_Milkano_-_Swordmasters_Youngest_Son.safetensors\", \"SD 1.5\", \"URL: https://civitai.com/models/2037308\\nTriggers: mesa milkano\\nBase Model: SD 1.5\\n\\nExamples:\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/6439eed5-1859-496f-8057-be09ba188650/original=true/105649114.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: breastplate, greaves, petals, looking_back, green_eyes, blunt_bangs, pink_hair, brown_belt, day, 1girl, bangs, night_sky, pants, arms_at_sides, star_\\\\(sky\\\\), gauntlets, sunset, sidelocks, blue_sky, night, belt, wariza, pov, black_pants, blush, vambraces, ponytail, katou_asuka\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e080074c-6edd-41d9-aa00-aad6068ef827/original=true/105649119.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: hands_on_own_knees, pov, black_gloves, bush, ponytail, open_mouth, mountain, brown_belt, floating_hair, solo, pink_hair, sidelocks, forest, looking_to_the_side, flower_field, crossed_legs, armored_boots, closed_mouth, petals, day, blunt_bangs, armor, field, long_hair, bangs, looking_back, night_sky, nature\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\\nOutput: https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/15419230-62c5-4ee7-9465-4b102e8a0071/original=true/105649111.jpeg\\nSize: 512x512\\nModel: 89927.safetensors\\nsteps: 30\\nhashes: {'model': '354B8C571D'}\\nprompt: nature, gloves, outdoors, ponytail, petals, looking_back, armor, , mountain, looking_at_viewer, forest, brown_belt, bush, hands_on_own_knees, alternate_hairstyle, tree, green_eyes, star_\\\\(sky\\\\), sky, sitting, grass, sunset, looking_to_the_side, bangs, cowboy_shot, blunt_bangs, smile, from_behind\\nsampler: DDIM\\ncfgScale: 7.5\\nresources: [{'hash': '354B8C571D', 'name': '89927.safetensors', 'type': 'model'}]\\nModel hash: 354B8C571D\\n\\n\"]}, {\"id\": 161, \"type\": \"LoraLoader\", \"pos\": [60.91288375854492, -547.70703125], \"size\": [315, 126], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 780}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 781}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [786], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [790], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"Orphia_-_Seirei_Gensouki.safetensors\", 0.6, 1]}, {\"id\": 159, \"type\": \"LoraLoader\", \"pos\": [55.07668685913086, -364.66888427734375], \"size\": [315, 126], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 786}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 790}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [772], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [787], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"age_slider_v20.safetensors\", 0, 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-317.33526611328125, -555.9864501953125], \"size\": [315, 98], \"flags\": {\"collapsed\": false}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [780], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [781], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"anyloracleanlinearmix_v10.safetensors\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [531.701904296875, -299.3091735839844], \"size\": [423.61407470703125, 228.15113830566406], \"flags\": {\"collapsed\": false}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 792}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \\n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \\n(indoor, dining room, sitting on chair, holding cup of tea:1.3), (looking at viewer, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [262, 3, 0, 8, 0, \"LATENT\"], [766, 158, 0, 157, 0, \"UPSCALE_MODEL\"], [767, 8, 0, 157, 1, \"IMAGE\"], [768, 157, 0, 9, 0, \"IMAGE\"], [772, 159, 0, 160, 0, \"MODEL\"], [775, 160, 1, 33, 0, \"CLIP\"], [780, 4, 0, 161, 0, \"MODEL\"], [781, 4, 1, 161, 1, \"CLIP\"], [786, 161, 0, 159, 0, \"MODEL\"], [787, 159, 1, 160, 1, \"CLIP\"], [790, 161, 1, 159, 1, \"CLIP\"], [791, 160, 0, 3, 0, \"MODEL\"], [792, 33, 0, 6, 0, \"CLIP\"], [793, 33, 0, 7, 0, \"CLIP\"], [794, 164, 0, 8, 1, \"VAE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1918176537727436, \"offset\": [293.1093270451885, 595.4901553635286]}, \"node_versions\": {\"comfy-core\": \"0.3.10\", \"lora-info\": \"5db14b8f37e65585d6114416feb8f7b110541548\"}}, \"version\": 0.4}}",
242
+ "steps": 10,
243
+ "width": 512,
244
+ "height": 768,
245
+ "models": [
246
+ "anyloracleanlinearmix_v10.safetensors"
247
+ ],
248
+ "prompt": "(orphia), (long_hair, breasts, blue_eyes, blonde_hair, bangs, hairband, green_hair, pointy_ears, medium_breasts, elf:1.32), \n(gloves, dress, ribbon, cleavage, bare_shoulders, hairband, fingerless_gloves, green_gloves, green_hairband:1.31), \n(indoor, dining room, sitting on chair, holding cup of tea:1.3), (looking at viewer, eye level shot:1.28), \n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)",
249
+ "denoise": 1,
250
+ "sampler": "LCM",
251
+ "cfgScale": 2,
252
+ "scheduler": "normal",
253
+ "upscalers": [
254
+ "4xUltrasharp_4xUltrasharpV10.pt"
255
+ ],
256
+ "negativePrompt": "embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), (unproportional), (text, brand, english text, korean text, text buble, logo:1.35)",
257
+ "additionalResources": [
258
+ {
259
+ "name": "PAseer-SD15-LCM Quick.safetensors",
260
+ "type": "lora",
261
+ "strength": 1,
262
+ "strengthClip": 1
263
+ },
264
+ {
265
+ "name": "Orphia_-_Seirei_Gensouki.safetensors",
266
+ "type": "lora",
267
+ "strength": 0.6,
268
+ "strengthClip": 1
269
+ }
270
+ ]
271
+ },
272
+ "availability": "Public",
273
+ "hasMeta": true,
274
+ "hasPositivePrompt": true,
275
+ "onSite": false,
276
+ "remixOfId": null
277
+ }
278
+ ],
279
+ "downloadUrl": "https://civitai.com/api/download/models/2312694",
280
+ "creator": {
281
+ "username": "MCKEMZIE",
282
+ "image": "https://lh3.googleusercontent.com/a/AAcHTtdwmJUl4IkaLPDRGhwpj6FlFEtQUF6NzneEhuLU3-Ek0AU=s96-c"
283
+ },
284
+ "extensions": {
285
+ "sd_civitai_helper": {
286
+ "version": "1.8.13",
287
+ "last_update": 1760877923,
288
+ "skeleton_file": false
289
+ }
290
+ }
291
+ }
A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.preview.png ADDED

Git LFS Details

  • SHA256: 9e48f8c9943fac7d1ce85cd2e1a2fa0626f99e4c20d9e93b1d95f24e14d6883d
  • Pointer size: 132 Bytes
  • Size of remote file: 6.36 MB
A-表番角色/精灵幻想记/Orphia_-_Seirei_Gensouki.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5c28c50f6b905d8b902af1d25282625efc46480e1586d72200af20a92319485
3
+ size 18991976
A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.civitai.info ADDED
@@ -0,0 +1,293 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2312455,
3
+ "modelId": 2043152,
4
+ "name": "V1",
5
+ "nsfwLevel": 2,
6
+ "createdAt": "2025-10-14T08:02:36.918Z",
7
+ "updatedAt": "2025-10-14T08:41:10.719Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-10-14T08:41:10.699Z",
10
+ "trainedWords": [
11
+ "sara"
12
+ ],
13
+ "trainingStatus": "Approved",
14
+ "trainingDetails": {
15
+ "type": "Character",
16
+ "params": {
17
+ "engine": "kohya",
18
+ "unetLR": 0.0005,
19
+ "clipSkip": 1,
20
+ "loraType": "lora",
21
+ "keepTokens": 0,
22
+ "networkDim": 32,
23
+ "numRepeats": 20,
24
+ "resolution": 512,
25
+ "lrScheduler": "cosine_with_restarts",
26
+ "minSnrGamma": 5,
27
+ "noiseOffset": 0.1,
28
+ "targetSteps": 500,
29
+ "enableBucket": true,
30
+ "networkAlpha": 16,
31
+ "optimizerType": "AdamW8Bit",
32
+ "textEncoderLR": 5e-05,
33
+ "maxTrainEpochs": 10,
34
+ "shuffleCaption": false,
35
+ "trainBatchSize": 6,
36
+ "flipAugmentation": false,
37
+ "lrSchedulerNumCycles": 3
38
+ },
39
+ "staging": false,
40
+ "baseModel": "sd_1_5",
41
+ "mediaType": "image",
42
+ "highPriority": false,
43
+ "baseModelType": "sd15",
44
+ "samplePrompts": [
45
+ "",
46
+ "",
47
+ ""
48
+ ],
49
+ "negativePrompt": "bad quality, low quality, worst quality, jpeg artifacts, blurry, pixelated, out of focus, watermark, text, signature"
50
+ },
51
+ "baseModel": "SD 1.5",
52
+ "baseModelType": "Standard",
53
+ "earlyAccessEndsAt": null,
54
+ "earlyAccessConfig": null,
55
+ "description": null,
56
+ "uploadType": "Trained",
57
+ "usageControl": "Download",
58
+ "air": "urn:air:sd1:lora:civitai:2043152@2312455",
59
+ "stats": {
60
+ "downloadCount": 36,
61
+ "ratingCount": 0,
62
+ "rating": 0,
63
+ "thumbsUpCount": 3
64
+ },
65
+ "model": {
66
+ "name": "Sara - Seirei Gensouki",
67
+ "type": "LORA",
68
+ "nsfw": false,
69
+ "poi": false,
70
+ "description": "Sara - Seirei Gensouki",
71
+ "tags": [
72
+ "character"
73
+ ],
74
+ "allowNoCredit": true,
75
+ "allowCommercialUse": [],
76
+ "allowDerivatives": true,
77
+ "allowDifferentLicense": true
78
+ },
79
+ "files": [
80
+ {
81
+ "id": 2202993,
82
+ "sizeKB": 36981.78125,
83
+ "name": "Sara_-_Seirei_Gensouki.safetensors",
84
+ "type": "Model",
85
+ "pickleScanResult": "Success",
86
+ "pickleScanMessage": "No Pickle imports",
87
+ "virusScanResult": "Success",
88
+ "virusScanMessage": null,
89
+ "scannedAt": "2025-10-14T08:20:43.914Z",
90
+ "metadata": {
91
+ "format": "SafeTensor"
92
+ },
93
+ "hashes": {
94
+ "AutoV1": "90C5D038",
95
+ "AutoV2": "D880C2CC03",
96
+ "SHA256": "D880C2CC031C460D346D8305C1CD4212CB823EBC969E42ADCDAC6693F3DC3495",
97
+ "CRC32": "FDA21DDB",
98
+ "BLAKE3": "CECF0FB0AD69B5549EFAE3F508F7A90452A11AF5484EA5C2598FBE5E991C803E",
99
+ "AutoV3": "272162238841"
100
+ },
101
+ "primary": true,
102
+ "downloadUrl": "https://civitai.com/api/download/models/2312455"
103
+ }
104
+ ],
105
+ "images": [
106
+ {
107
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/18ce0fcb-956f-4ff1-a521-b9671ce4f232/original=true/105999667.jpeg",
108
+ "nsfwLevel": 2,
109
+ "width": 2048,
110
+ "height": 3072,
111
+ "hash": "U3FYcbHr00X1K*Ri01oH00=|I_oM00ad}[56",
112
+ "type": "image",
113
+ "metadata": {
114
+ "hash": "U3FYcbHr00X1K*Ri01oH00=|I_oM00ad}[56",
115
+ "size": 7666664,
116
+ "width": 2048,
117
+ "height": 3072
118
+ },
119
+ "minor": false,
120
+ "poi": false,
121
+ "meta": {
122
+ "seed": 408136856004466,
123
+ "vaes": [
124
+ "CleanVAE.safetensors"
125
+ ],
126
+ "Model": "cleanlinearmix_nsfw",
127
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 408136856004466, \"steps\": 10, \"cfg\": 2.0, \"sampler_name\": \"lcm\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"160\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"cleanlinearmix_nsfw.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 1}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \\n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \\n(standing, dynamic pose:1.25), \\n(outdoor, at night, starry night:1.3), (looking at viewer, front view, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"164\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"157\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"33\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"160\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}, \"157\": {\"inputs\": {\"upscale_model\": [\"158\", 0], \"image\": [\"8\", 0]}, \"class_type\": \"ImageUpscaleWithModel\", \"_meta\": {\"title\": \"Upscale Image (using Model)\"}}, \"158\": {\"inputs\": {\"model_name\": \"4xUltrasharp_4xUltrasharpV10.pt\"}, \"class_type\": \"UpscaleModelLoader\", \"_meta\": {\"title\": \"Load Upscale Model\"}}, \"159\": {\"inputs\": {\"lora_name\": \"age_slider_v20.safetensors\", \"strength_model\": 0.0, \"strength_clip\": 1.0, \"model\": [\"161\", 0], \"clip\": [\"161\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"160\": {\"inputs\": {\"lora_name\": \"PAseer-SD15-LCM Quick.safetensors\", \"strength_model\": 1.0, \"strength_clip\": 1.0, \"model\": [\"159\", 0], \"clip\": [\"159\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"161\": {\"inputs\": {\"lora_name\": \"Sara_-_Seirei_Gensouki.safetensors\", \"strength_model\": 0.75, \"strength_clip\": 1.0, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"163\": {\"inputs\": {\"lora_name\": \"Sara_-_Seirei_Gensouki.safetensors\", \"Base Model\": \"\", \"output\": \"\"}, \"class_type\": \"LoraInfo\", \"_meta\": {\"title\": \"Lora Info\"}}, \"164\": {\"inputs\": {\"vae_name\": \"CleanVAE.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}}, \"workflow\": {\"last_node_id\": 164, \"last_link_id\": 794, \"nodes\": [{\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [629.10693359375, 235.40773010253906], \"size\": [315, 106], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [512, 768, 1]}, {\"id\": 157, \"type\": \"ImageUpscaleWithModel\", \"pos\": [1911.0567626953125, -203.64340209960938], \"size\": [340.20001220703125, 46], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"upscale_model\", \"type\": \"UPSCALE_MODEL\", \"link\": 766}, {\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 767}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [768], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"ImageUpscaleWithModel\"}, \"widgets_values\": []}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [2322.21484375, -169.14207458496094], \"size\": [364.16668701171875, 404.3333740234375], \"flags\": {\"collapsed\": true}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 768}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1580.349365234375, -17.497373580932617], \"size\": [210, 46], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 262}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 794}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [767], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 158, \"type\": \"UpscaleModelLoader\", \"pos\": [1537.1036376953125, -273.8653259277344], \"size\": [315, 58], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"UPSCALE_MODEL\", \"type\": \"UPSCALE_MODEL\", \"links\": [766], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UpscaleModelLoader\"}, \"widgets_values\": [\"4xUltrasharp_4xUltrasharpV10.pt\"]}, {\"id\": 33, \"type\": \"CLIPSetLastLayer\", \"pos\": [66.87496948242188, 47.42504119873047], \"size\": [315, 58], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 775}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [792, 793], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPSetLastLayer\"}, \"widgets_values\": [-2]}, {\"id\": 164, \"type\": \"VAELoader\", \"pos\": [1144.0006103515625, 76.65125274658203], \"size\": [315, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [794], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"CleanVAE.safetensors\"]}, {\"id\": 160, \"type\": \"LoraLoader\", \"pos\": [58.979976654052734, -166.80343627929688], \"size\": [315, 126], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 772}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 787}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [791], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [775], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"PAseer-SD15-LCM Quick.safetensors\", 1, 1]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [1130.9888916015625, -342.4229431152344], \"size\": [315, 262], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 791}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [262], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [408136856004466, \"randomize\", 10, 2, \"lcm\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [528.8944091796875, -10.626336097717285], \"size\": [425.27801513671875, 180.6060791015625], \"flags\": {\"collapsed\": false}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 793}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)\"]}, {\"id\": 159, \"type\": \"LoraLoader\", \"pos\": [55.07668685913086, -364.66888427734375], \"size\": [315, 126], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 786}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 790}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [772], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [787], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"age_slider_v20.safetensors\", 0, 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-317.33526611328125, -555.9864501953125], \"size\": [315, 98], \"flags\": {\"collapsed\": false}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [780], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [781], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cleanlinearmix_nsfw.safetensors\"]}, {\"id\": 163, \"type\": \"LoraInfo\", \"pos\": [534.1552734375, -668.4848022460938], \"size\": [346.52117919921875, 293.06060791015625], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"lora_name\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"trigger_words\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"example_prompt\", \"type\": \"STRING\", \"links\": null}], \"properties\": {\"Node name for S&R\": \"LoraInfo\"}, \"widgets_values\": [\"Sara_-_Seirei_Gensouki.safetensors\", \"\", \"\"]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [531.701904296875, -299.3091735839844], \"size\": [423.61407470703125, 228.15113830566406], \"flags\": {\"collapsed\": false}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 792}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \\n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \\n(standing, dynamic pose:1.25), \\n(outdoor, at night, starry night:1.3), (looking at viewer, front view, eye level shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\"]}, {\"id\": 161, \"type\": \"LoraLoader\", \"pos\": [60.91288375854492, -547.70703125], \"size\": [315, 126], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 780}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 781}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [786], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [790], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"Sara_-_Seirei_Gensouki.safetensors\", 0.75, 1]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [262, 3, 0, 8, 0, \"LATENT\"], [766, 158, 0, 157, 0, \"UPSCALE_MODEL\"], [767, 8, 0, 157, 1, \"IMAGE\"], [768, 157, 0, 9, 0, \"IMAGE\"], [772, 159, 0, 160, 0, \"MODEL\"], [775, 160, 1, 33, 0, \"CLIP\"], [780, 4, 0, 161, 0, \"MODEL\"], [781, 4, 1, 161, 1, \"CLIP\"], [786, 161, 0, 159, 0, \"MODEL\"], [787, 159, 1, 160, 1, \"CLIP\"], [790, 161, 1, 159, 1, \"CLIP\"], [791, 160, 0, 3, 0, \"MODEL\"], [792, 33, 0, 6, 0, \"CLIP\"], [793, 33, 0, 7, 0, \"CLIP\"], [794, 164, 0, 8, 1, \"VAE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1918176537727436, \"offset\": [-33.42264175849948, 562.6271564420714]}, \"node_versions\": {\"comfy-core\": \"0.3.10\", \"lora-info\": \"5db14b8f37e65585d6114416feb8f7b110541548\"}}, \"version\": 0.4}}",
128
+ "steps": 10,
129
+ "width": 512,
130
+ "height": 768,
131
+ "models": [
132
+ "cleanlinearmix_nsfw.safetensors"
133
+ ],
134
+ "prompt": "(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \n(standing, dynamic pose:1.25), \n(outdoor, at night, starry night:1.3), (looking at viewer, front view, eye level shot:1.28), \n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)",
135
+ "denoise": 1,
136
+ "sampler": "LCM",
137
+ "cfgScale": 2,
138
+ "scheduler": "normal",
139
+ "upscalers": [
140
+ "4xUltrasharp_4xUltrasharpV10.pt"
141
+ ],
142
+ "negativePrompt": "embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)",
143
+ "additionalResources": [
144
+ {
145
+ "name": "PAseer-SD15-LCM Quick.safetensors",
146
+ "type": "lora",
147
+ "strength": 1,
148
+ "strengthClip": 1
149
+ },
150
+ {
151
+ "name": "Sara_-_Seirei_Gensouki.safetensors",
152
+ "type": "lora",
153
+ "strength": 0.75,
154
+ "strengthClip": 1
155
+ }
156
+ ]
157
+ },
158
+ "availability": "Public",
159
+ "hasMeta": true,
160
+ "hasPositivePrompt": true,
161
+ "onSite": false,
162
+ "remixOfId": null
163
+ },
164
+ {
165
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/e5836c86-1db4-43db-904f-287580b10d66/original=true/105999668.jpeg",
166
+ "nsfwLevel": 2,
167
+ "width": 2048,
168
+ "height": 3072,
169
+ "hash": "U8Frbd9E00?F01^%WF4nI8In?Ht759jb~Bog",
170
+ "type": "image",
171
+ "metadata": {
172
+ "hash": "U8Frbd9E00?F01^%WF4nI8In?Ht759jb~Bog",
173
+ "size": 7112967,
174
+ "width": 2048,
175
+ "height": 3072
176
+ },
177
+ "minor": false,
178
+ "poi": false,
179
+ "meta": {
180
+ "seed": 159772702462007,
181
+ "vaes": [
182
+ "CleanVAE.safetensors"
183
+ ],
184
+ "Model": "cleanlinearmix_nsfw",
185
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 159772702462007, \"steps\": 10, \"cfg\": 2.0, \"sampler_name\": \"lcm\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"160\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"cleanlinearmix_nsfw.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 1}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \\n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \\n(sitting:1.25), \\n(outdoor, at night, starry night:1.3), (looking at viewer, front view, dramatic shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"164\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"157\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"33\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"160\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}, \"157\": {\"inputs\": {\"upscale_model\": [\"158\", 0], \"image\": [\"8\", 0]}, \"class_type\": \"ImageUpscaleWithModel\", \"_meta\": {\"title\": \"Upscale Image (using Model)\"}}, \"158\": {\"inputs\": {\"model_name\": \"4xUltrasharp_4xUltrasharpV10.pt\"}, \"class_type\": \"UpscaleModelLoader\", \"_meta\": {\"title\": \"Load Upscale Model\"}}, \"159\": {\"inputs\": {\"lora_name\": \"age_slider_v20.safetensors\", \"strength_model\": 0.0, \"strength_clip\": 1.0, \"model\": [\"161\", 0], \"clip\": [\"161\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"160\": {\"inputs\": {\"lora_name\": \"PAseer-SD15-LCM Quick.safetensors\", \"strength_model\": 1.0, \"strength_clip\": 1.0, \"model\": [\"159\", 0], \"clip\": [\"159\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"161\": {\"inputs\": {\"lora_name\": \"Sara_-_Seirei_Gensouki.safetensors\", \"strength_model\": 0.75, \"strength_clip\": 1.0, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"163\": {\"inputs\": {\"lora_name\": \"Sara_-_Seirei_Gensouki.safetensors\", \"Base Model\": \"\", \"output\": \"\"}, \"class_type\": \"LoraInfo\", \"_meta\": {\"title\": \"Lora Info\"}}, \"164\": {\"inputs\": {\"vae_name\": \"CleanVAE.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}}, \"workflow\": {\"last_node_id\": 164, \"last_link_id\": 794, \"nodes\": [{\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [629.10693359375, 235.40773010253906], \"size\": [315, 106], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [512, 768, 1]}, {\"id\": 157, \"type\": \"ImageUpscaleWithModel\", \"pos\": [1911.0567626953125, -203.64340209960938], \"size\": [340.20001220703125, 46], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"upscale_model\", \"type\": \"UPSCALE_MODEL\", \"link\": 766}, {\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 767}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [768], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"ImageUpscaleWithModel\"}, \"widgets_values\": []}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [2322.21484375, -169.14207458496094], \"size\": [364.16668701171875, 404.3333740234375], \"flags\": {\"collapsed\": true}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 768}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1580.349365234375, -17.497373580932617], \"size\": [210, 46], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 262}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 794}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [767], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 158, \"type\": \"UpscaleModelLoader\", \"pos\": [1537.1036376953125, -273.8653259277344], \"size\": [315, 58], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"UPSCALE_MODEL\", \"type\": \"UPSCALE_MODEL\", \"links\": [766], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UpscaleModelLoader\"}, \"widgets_values\": [\"4xUltrasharp_4xUltrasharpV10.pt\"]}, {\"id\": 33, \"type\": \"CLIPSetLastLayer\", \"pos\": [66.87496948242188, 47.42504119873047], \"size\": [315, 58], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 775}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [792, 793], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPSetLastLayer\"}, \"widgets_values\": [-2]}, {\"id\": 164, \"type\": \"VAELoader\", \"pos\": [1144.0006103515625, 76.65125274658203], \"size\": [315, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [794], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"CleanVAE.safetensors\"]}, {\"id\": 160, \"type\": \"LoraLoader\", \"pos\": [58.979976654052734, -166.80343627929688], \"size\": [315, 126], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 772}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 787}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [791], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [775], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"PAseer-SD15-LCM Quick.safetensors\", 1, 1]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [1130.9888916015625, -342.4229431152344], \"size\": [315, 262], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 791}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [262], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [159772702462007, \"randomize\", 10, 2, \"lcm\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [528.8944091796875, -10.626336097717285], \"size\": [425.27801513671875, 180.6060791015625], \"flags\": {\"collapsed\": false}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 793}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)\"]}, {\"id\": 159, \"type\": \"LoraLoader\", \"pos\": [55.07668685913086, -364.66888427734375], \"size\": [315, 126], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 786}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 790}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [772], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [787], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"age_slider_v20.safetensors\", 0, 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-317.33526611328125, -555.9864501953125], \"size\": [315, 98], \"flags\": {\"collapsed\": false}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [780], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [781], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cleanlinearmix_nsfw.safetensors\"]}, {\"id\": 163, \"type\": \"LoraInfo\", \"pos\": [534.1552734375, -668.4848022460938], \"size\": [346.52117919921875, 293.06060791015625], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"lora_name\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"trigger_words\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"example_prompt\", \"type\": \"STRING\", \"links\": null}], \"properties\": {\"Node name for S&R\": \"LoraInfo\"}, \"widgets_values\": [\"Sara_-_Seirei_Gensouki.safetensors\", \"\", \"\"]}, {\"id\": 161, \"type\": \"LoraLoader\", \"pos\": [60.91288375854492, -547.70703125], \"size\": [315, 126], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 780}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 781}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [786], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [790], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"Sara_-_Seirei_Gensouki.safetensors\", 0.75, 1]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [531.701904296875, -299.3091735839844], \"size\": [423.61407470703125, 228.15113830566406], \"flags\": {\"collapsed\": false}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 792}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \\n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \\n(sitting:1.25), \\n(outdoor, at night, starry night:1.3), (looking at viewer, front view, dramatic shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [262, 3, 0, 8, 0, \"LATENT\"], [766, 158, 0, 157, 0, \"UPSCALE_MODEL\"], [767, 8, 0, 157, 1, \"IMAGE\"], [768, 157, 0, 9, 0, \"IMAGE\"], [772, 159, 0, 160, 0, \"MODEL\"], [775, 160, 1, 33, 0, \"CLIP\"], [780, 4, 0, 161, 0, \"MODEL\"], [781, 4, 1, 161, 1, \"CLIP\"], [786, 161, 0, 159, 0, \"MODEL\"], [787, 159, 1, 160, 1, \"CLIP\"], [790, 161, 1, 159, 1, \"CLIP\"], [791, 160, 0, 3, 0, \"MODEL\"], [792, 33, 0, 6, 0, \"CLIP\"], [793, 33, 0, 7, 0, \"CLIP\"], [794, 164, 0, 8, 1, \"VAE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1918176537727436, \"offset\": [-33.42264175849948, 562.6271564420714]}, \"node_versions\": {\"comfy-core\": \"0.3.10\", \"lora-info\": \"5db14b8f37e65585d6114416feb8f7b110541548\"}}, \"version\": 0.4}}",
186
+ "steps": 10,
187
+ "width": 512,
188
+ "height": 768,
189
+ "models": [
190
+ "cleanlinearmix_nsfw.safetensors"
191
+ ],
192
+ "prompt": "(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \n(sitting:1.25), \n(outdoor, at night, starry night:1.3), (looking at viewer, front view, dramatic shot:1.28), \n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)",
193
+ "denoise": 1,
194
+ "sampler": "LCM",
195
+ "cfgScale": 2,
196
+ "scheduler": "normal",
197
+ "upscalers": [
198
+ "4xUltrasharp_4xUltrasharpV10.pt"
199
+ ],
200
+ "negativePrompt": "embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)",
201
+ "additionalResources": [
202
+ {
203
+ "name": "PAseer-SD15-LCM Quick.safetensors",
204
+ "type": "lora",
205
+ "strength": 1,
206
+ "strengthClip": 1
207
+ },
208
+ {
209
+ "name": "Sara_-_Seirei_Gensouki.safetensors",
210
+ "type": "lora",
211
+ "strength": 0.75,
212
+ "strengthClip": 1
213
+ }
214
+ ]
215
+ },
216
+ "availability": "Public",
217
+ "hasMeta": true,
218
+ "hasPositivePrompt": true,
219
+ "onSite": false,
220
+ "remixOfId": null
221
+ },
222
+ {
223
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/0ae7647c-8eaf-4467-9482-b1e8699c7d0e/original=true/105999666.jpeg",
224
+ "nsfwLevel": 2,
225
+ "width": 2048,
226
+ "height": 3072,
227
+ "hash": "U7F=tENM00?G9tt69ZIU00%1IWE19FM{~Ut6",
228
+ "type": "image",
229
+ "metadata": {
230
+ "hash": "U7F=tENM00?G9tt69ZIU00%1IWE19FM{~Ut6",
231
+ "size": 7453277,
232
+ "width": 2048,
233
+ "height": 3072
234
+ },
235
+ "minor": false,
236
+ "poi": false,
237
+ "meta": {
238
+ "seed": 982705919184787,
239
+ "vaes": [
240
+ "CleanVAE.safetensors"
241
+ ],
242
+ "Model": "cleanlinearmix_nsfw",
243
+ "comfy": "{\"prompt\": {\"3\": {\"inputs\": {\"seed\": 982705919184787, \"steps\": 10, \"cfg\": 2.0, \"sampler_name\": \"lcm\", \"scheduler\": \"normal\", \"denoise\": 1.0, \"model\": [\"160\", 0], \"positive\": [\"6\", 0], \"negative\": [\"7\", 0], \"latent_image\": [\"5\", 0]}, \"class_type\": \"KSampler\", \"_meta\": {\"title\": \"KSampler\"}}, \"4\": {\"inputs\": {\"ckpt_name\": \"cleanlinearmix_nsfw.safetensors\"}, \"class_type\": \"CheckpointLoaderSimple\", \"_meta\": {\"title\": \"Load Checkpoint\"}}, \"5\": {\"inputs\": {\"width\": 512, \"height\": 768, \"batch_size\": 1}, \"class_type\": \"EmptyLatentImage\", \"_meta\": {\"title\": \"Empty Latent Image\"}}, \"6\": {\"inputs\": {\"text\": \"(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \\n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \\n(standing:1.25), \\n(outdoor, at night, starry night:1.3), (looking at viewer, front view, cowboy shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"7\": {\"inputs\": {\"text\": \"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)\", \"clip\": [\"33\", 0]}, \"class_type\": \"CLIPTextEncode\", \"_meta\": {\"title\": \"CLIP Text Encode (Prompt)\"}}, \"8\": {\"inputs\": {\"samples\": [\"3\", 0], \"vae\": [\"164\", 0]}, \"class_type\": \"VAEDecode\", \"_meta\": {\"title\": \"VAE Decode\"}}, \"9\": {\"inputs\": {\"filename_prefix\": \"ComfyUI\", \"images\": [\"157\", 0]}, \"class_type\": \"SaveImage\", \"_meta\": {\"title\": \"Save Image\"}}, \"33\": {\"inputs\": {\"stop_at_clip_layer\": -2, \"clip\": [\"160\", 1]}, \"class_type\": \"CLIPSetLastLayer\", \"_meta\": {\"title\": \"CLIP Set Last Layer\"}}, \"157\": {\"inputs\": {\"upscale_model\": [\"158\", 0], \"image\": [\"8\", 0]}, \"class_type\": \"ImageUpscaleWithModel\", \"_meta\": {\"title\": \"Upscale Image (using Model)\"}}, \"158\": {\"inputs\": {\"model_name\": \"4xUltrasharp_4xUltrasharpV10.pt\"}, \"class_type\": \"UpscaleModelLoader\", \"_meta\": {\"title\": \"Load Upscale Model\"}}, \"159\": {\"inputs\": {\"lora_name\": \"age_slider_v20.safetensors\", \"strength_model\": 0.0, \"strength_clip\": 1.0, \"model\": [\"161\", 0], \"clip\": [\"161\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"160\": {\"inputs\": {\"lora_name\": \"PAseer-SD15-LCM Quick.safetensors\", \"strength_model\": 1.0, \"strength_clip\": 1.0, \"model\": [\"159\", 0], \"clip\": [\"159\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"161\": {\"inputs\": {\"lora_name\": \"Sara_-_Seirei_Gensouki.safetensors\", \"strength_model\": 0.6, \"strength_clip\": 1.0, \"model\": [\"4\", 0], \"clip\": [\"4\", 1]}, \"class_type\": \"LoraLoader\", \"_meta\": {\"title\": \"Load LoRA\"}}, \"163\": {\"inputs\": {\"lora_name\": \"Sara_-_Seirei_Gensouki.safetensors\", \"Base Model\": \"\", \"output\": \"\"}, \"class_type\": \"LoraInfo\", \"_meta\": {\"title\": \"Lora Info\"}}, \"164\": {\"inputs\": {\"vae_name\": \"CleanVAE.safetensors\"}, \"class_type\": \"VAELoader\", \"_meta\": {\"title\": \"Load VAE\"}}}, \"workflow\": {\"last_node_id\": 164, \"last_link_id\": 794, \"nodes\": [{\"id\": 5, \"type\": \"EmptyLatentImage\", \"pos\": [629.10693359375, 235.40773010253906], \"size\": [315, 106], \"flags\": {}, \"order\": 0, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [2], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"EmptyLatentImage\"}, \"widgets_values\": [512, 768, 1]}, {\"id\": 157, \"type\": \"ImageUpscaleWithModel\", \"pos\": [1911.0567626953125, -203.64340209960938], \"size\": [340.20001220703125, 46], \"flags\": {}, \"order\": 13, \"mode\": 0, \"inputs\": [{\"name\": \"upscale_model\", \"type\": \"UPSCALE_MODEL\", \"link\": 766}, {\"name\": \"image\", \"type\": \"IMAGE\", \"link\": 767}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [768], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"ImageUpscaleWithModel\"}, \"widgets_values\": []}, {\"id\": 9, \"type\": \"SaveImage\", \"pos\": [2322.21484375, -169.14207458496094], \"size\": [364.16668701171875, 404.3333740234375], \"flags\": {\"collapsed\": true}, \"order\": 14, \"mode\": 0, \"inputs\": [{\"name\": \"images\", \"type\": \"IMAGE\", \"link\": 768}], \"outputs\": [], \"properties\": {}, \"widgets_values\": [\"ComfyUI\"]}, {\"id\": 8, \"type\": \"VAEDecode\", \"pos\": [1580.349365234375, -17.497373580932617], \"size\": [210, 46], \"flags\": {}, \"order\": 12, \"mode\": 0, \"inputs\": [{\"name\": \"samples\", \"type\": \"LATENT\", \"link\": 262}, {\"name\": \"vae\", \"type\": \"VAE\", \"link\": 794}], \"outputs\": [{\"name\": \"IMAGE\", \"type\": \"IMAGE\", \"links\": [767], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAEDecode\"}, \"widgets_values\": []}, {\"id\": 158, \"type\": \"UpscaleModelLoader\", \"pos\": [1537.1036376953125, -273.8653259277344], \"size\": [315, 58], \"flags\": {}, \"order\": 1, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"UPSCALE_MODEL\", \"type\": \"UPSCALE_MODEL\", \"links\": [766], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"UpscaleModelLoader\"}, \"widgets_values\": [\"4xUltrasharp_4xUltrasharpV10.pt\"]}, {\"id\": 33, \"type\": \"CLIPSetLastLayer\", \"pos\": [66.87496948242188, 47.42504119873047], \"size\": [315, 58], \"flags\": {}, \"order\": 8, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 775}], \"outputs\": [{\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [792, 793], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPSetLastLayer\"}, \"widgets_values\": [-2]}, {\"id\": 164, \"type\": \"VAELoader\", \"pos\": [1144.0006103515625, 76.65125274658203], \"size\": [315, 58], \"flags\": {}, \"order\": 2, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [794], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"VAELoader\"}, \"widgets_values\": [\"CleanVAE.safetensors\"]}, {\"id\": 160, \"type\": \"LoraLoader\", \"pos\": [58.979976654052734, -166.80343627929688], \"size\": [315, 126], \"flags\": {}, \"order\": 7, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 772}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 787}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [791], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [775], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"PAseer-SD15-LCM Quick.safetensors\", 1, 1]}, {\"id\": 3, \"type\": \"KSampler\", \"pos\": [1130.9888916015625, -342.4229431152344], \"size\": [315, 262], \"flags\": {}, \"order\": 11, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 791}, {\"name\": \"positive\", \"type\": \"CONDITIONING\", \"link\": 4}, {\"name\": \"negative\", \"type\": \"CONDITIONING\", \"link\": 6}, {\"name\": \"latent_image\", \"type\": \"LATENT\", \"link\": 2}], \"outputs\": [{\"name\": \"LATENT\", \"type\": \"LATENT\", \"links\": [262], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"KSampler\"}, \"widgets_values\": [982705919184787, \"randomize\", 10, 2, \"lcm\", \"normal\", 1]}, {\"id\": 7, \"type\": \"CLIPTextEncode\", \"pos\": [528.8944091796875, -10.626336097717285], \"size\": [425.27801513671875, 180.6060791015625], \"flags\": {\"collapsed\": false}, \"order\": 10, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 793}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [6], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)\"]}, {\"id\": 159, \"type\": \"LoraLoader\", \"pos\": [55.07668685913086, -364.66888427734375], \"size\": [315, 126], \"flags\": {}, \"order\": 6, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 786}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 790}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [772], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [787], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"age_slider_v20.safetensors\", 0, 1]}, {\"id\": 4, \"type\": \"CheckpointLoaderSimple\", \"pos\": [-317.33526611328125, -555.9864501953125], \"size\": [315, 98], \"flags\": {\"collapsed\": false}, \"order\": 3, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [780], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [781], \"slot_index\": 1}, {\"name\": \"VAE\", \"type\": \"VAE\", \"links\": [], \"slot_index\": 2}], \"properties\": {\"Node name for S&R\": \"CheckpointLoaderSimple\"}, \"widgets_values\": [\"cleanlinearmix_nsfw.safetensors\"]}, {\"id\": 163, \"type\": \"LoraInfo\", \"pos\": [534.1552734375, -668.4848022460938], \"size\": [346.52117919921875, 293.06060791015625], \"flags\": {}, \"order\": 4, \"mode\": 0, \"inputs\": [], \"outputs\": [{\"name\": \"lora_name\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"trigger_words\", \"type\": \"STRING\", \"links\": null}, {\"name\": \"example_prompt\", \"type\": \"STRING\", \"links\": null}], \"properties\": {\"Node name for S&R\": \"LoraInfo\"}, \"widgets_values\": [\"Sara_-_Seirei_Gensouki.safetensors\", \"\", \"\"]}, {\"id\": 161, \"type\": \"LoraLoader\", \"pos\": [60.91288375854492, -547.70703125], \"size\": [315, 126], \"flags\": {}, \"order\": 5, \"mode\": 0, \"inputs\": [{\"name\": \"model\", \"type\": \"MODEL\", \"link\": 780}, {\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 781}], \"outputs\": [{\"name\": \"MODEL\", \"type\": \"MODEL\", \"links\": [786], \"slot_index\": 0}, {\"name\": \"CLIP\", \"type\": \"CLIP\", \"links\": [790], \"slot_index\": 1}], \"properties\": {\"Node name for S&R\": \"LoraLoader\"}, \"widgets_values\": [\"Sara_-_Seirei_Gensouki.safetensors\", 0.6, 1]}, {\"id\": 6, \"type\": \"CLIPTextEncode\", \"pos\": [531.701904296875, -299.3091735839844], \"size\": [423.61407470703125, 228.15113830566406], \"flags\": {\"collapsed\": false}, \"order\": 9, \"mode\": 0, \"inputs\": [{\"name\": \"clip\", \"type\": \"CLIP\", \"link\": 792}], \"outputs\": [{\"name\": \"CONDITIONING\", \"type\": \"CONDITIONING\", \"links\": [4], \"slot_index\": 0}], \"properties\": {\"Node name for S&R\": \"CLIPTextEncode\"}, \"widgets_values\": [\"(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \\n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \\n(standing:1.25), \\n(outdoor, at night, starry night:1.3), (looking at viewer, front view, cowboy shot:1.28), \\n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)\"]}], \"links\": [[2, 5, 0, 3, 3, \"LATENT\"], [4, 6, 0, 3, 1, \"CONDITIONING\"], [6, 7, 0, 3, 2, \"CONDITIONING\"], [262, 3, 0, 8, 0, \"LATENT\"], [766, 158, 0, 157, 0, \"UPSCALE_MODEL\"], [767, 8, 0, 157, 1, \"IMAGE\"], [768, 157, 0, 9, 0, \"IMAGE\"], [772, 159, 0, 160, 0, \"MODEL\"], [775, 160, 1, 33, 0, \"CLIP\"], [780, 4, 0, 161, 0, \"MODEL\"], [781, 4, 1, 161, 1, \"CLIP\"], [786, 161, 0, 159, 0, \"MODEL\"], [787, 159, 1, 160, 1, \"CLIP\"], [790, 161, 1, 159, 1, \"CLIP\"], [791, 160, 0, 3, 0, \"MODEL\"], [792, 33, 0, 6, 0, \"CLIP\"], [793, 33, 0, 7, 0, \"CLIP\"], [794, 164, 0, 8, 1, \"VAE\"]], \"groups\": [], \"config\": {}, \"extra\": {\"ds\": {\"scale\": 1.1918176537727436, \"offset\": [95.93156396965216, 503.1941315805541]}, \"node_versions\": {\"comfy-core\": \"0.3.10\", \"lora-info\": \"5db14b8f37e65585d6114416feb8f7b110541548\"}}, \"version\": 0.4}}",
244
+ "steps": 10,
245
+ "width": 512,
246
+ "height": 768,
247
+ "models": [
248
+ "cleanlinearmix_nsfw.safetensors"
249
+ ],
250
+ "prompt": "(1girl, solo), (sara), (bangs, smile, animal_ears, twintails, yellow_eyes, grey_hair, very_long_hair, collarbone, animal_ear_fluff:1.32), \n(wear skirt, shirt, gloves, bow, bare_shoulders, short_sleeves, fingerless_gloves, off_shoulder, blue_skirt, crop_top, blue_shirt, purple_bow, off-shoulder_shirt:1.31), \n(standing:1.25), \n(outdoor, at night, starry night:1.3), (looking at viewer, front view, cowboy shot:1.28), \n(best quality, masterpiece:1.26), high resolution, unity 8k wallpaper, (illustration:0.8), (beautiful detailed eyes:1.1), extremely detailed face, (perfect lighting), extremely detailed CG, (perfect hands, perfect anatomy)",
251
+ "denoise": 1,
252
+ "sampler": "LCM",
253
+ "cfgScale": 2,
254
+ "scheduler": "normal",
255
+ "upscalers": [
256
+ "4xUltrasharp_4xUltrasharpV10.pt"
257
+ ],
258
+ "negativePrompt": "embedding:EasyNegative, embedding:verybadimagenegative_v1.3, embedding:ng_deepnegative_v1_75t, embedding:bad-hands-5, (black and white, B&W:1.5), (bad quality, worst quality:1.3),worst detail,sketch,censor, 2girls, (loli, child, chibi:1.35), unproportional, (text, brand, english text, korean text, text buble, logo:1.35), (muscular)",
259
+ "additionalResources": [
260
+ {
261
+ "name": "PAseer-SD15-LCM Quick.safetensors",
262
+ "type": "lora",
263
+ "strength": 1,
264
+ "strengthClip": 1
265
+ },
266
+ {
267
+ "name": "Sara_-_Seirei_Gensouki.safetensors",
268
+ "type": "lora",
269
+ "strength": 0.6,
270
+ "strengthClip": 1
271
+ }
272
+ ]
273
+ },
274
+ "availability": "Public",
275
+ "hasMeta": true,
276
+ "hasPositivePrompt": true,
277
+ "onSite": false,
278
+ "remixOfId": null
279
+ }
280
+ ],
281
+ "downloadUrl": "https://civitai.com/api/download/models/2312455",
282
+ "creator": {
283
+ "username": "MCKEMZIE",
284
+ "image": "https://lh3.googleusercontent.com/a/AAcHTtdwmJUl4IkaLPDRGhwpj6FlFEtQUF6NzneEhuLU3-Ek0AU=s96-c"
285
+ },
286
+ "extensions": {
287
+ "sd_civitai_helper": {
288
+ "version": "1.8.13",
289
+ "last_update": 1760877979,
290
+ "skeleton_file": false
291
+ }
292
+ }
293
+ }
A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.preview.png ADDED

Git LFS Details

  • SHA256: 1b7b1c872a802cc148a06c60ffd12b31744cd0ec7dbdee0e9ac8b5a5da9b23cd
  • Pointer size: 132 Bytes
  • Size of remote file: 7.67 MB
A-表番角色/精灵幻想记/Sara_-_Seirei_Gensouki.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d880c2cc031c460d346d8305c1cd4212cb823ebc969e42adcdac6693f3dc3495
3
+ size 37869344
A-表番角色/精灵幻想记/aishia.preview.png CHANGED

Git LFS Details

  • SHA256: e0def2e7f8f75d644b9a2b62d62711abf851a4b38e0d859f3120f09d0b983351
  • Pointer size: 131 Bytes
  • Size of remote file: 166 kB
A-表番角色/精灵幻想记/dryas.preview.png CHANGED

Git LFS Details

  • SHA256: 0f47c4d7cbefa9ce5ebff1e4e9b0382e085986010779047118bd0f568c4987d2
  • Pointer size: 131 Bytes
  • Size of remote file: 148 kB
A-表番角色/精灵幻想记/orphia.preview.png CHANGED

Git LFS Details

  • SHA256: 8a5d3850b3e36149e2662328a35533db4f1c6f099daa78f0e9efbc7fcbc9105e
  • Pointer size: 131 Bytes
  • Size of remote file: 175 kB
A-表番角色/精灵幻想记/orphia_no_outfit.preview.png CHANGED

Git LFS Details

  • SHA256: 910d4a5a0dfa1fa4f4c4b3714a7e7d1ca8ef4a2b36429e64b192d7a2c6af7a3e
  • Pointer size: 131 Bytes
  • Size of remote file: 130 kB
A-表番角色/精灵幻想记/精霊幻想記莎拉(サラ).preview.png CHANGED

Git LFS Details

  • SHA256: 14ac5515300039d32b7033a65fb5ed4247701ecd4475197107176d50efbe9d39
  • Pointer size: 131 Bytes
  • Size of remote file: 477 kB
A-表番角色/高达系列/AmiaLee-10.civitai.info ADDED
@@ -0,0 +1,179 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "id": 2231518,
3
+ "modelId": 1971514,
4
+ "name": "v1.0",
5
+ "nsfwLevel": 2,
6
+ "createdAt": "2025-09-19T16:04:05.889Z",
7
+ "updatedAt": "2025-09-19T16:05:55.768Z",
8
+ "status": "Published",
9
+ "publishedAt": "2025-09-19T16:05:55.747Z",
10
+ "trainedWords": [
11
+ "AmiaLee, 1girl, solo, short hair, white shirt, pleated skirt, brown hair, long sleeves, black thighhighs, neck ribbon, brown eyes, brown blazer, school uniform, pleated skirt, miniskirt, open blazer, blue skirt, neck ribbon, bob cut, yellow sweater vest"
12
+ ],
13
+ "trainingStatus": null,
14
+ "trainingDetails": null,
15
+ "baseModel": "SD 1.5",
16
+ "baseModelType": "Standard",
17
+ "earlyAccessEndsAt": null,
18
+ "earlyAccessConfig": null,
19
+ "description": null,
20
+ "uploadType": "Created",
21
+ "usageControl": "Download",
22
+ "air": "urn:air:sd1:lora:civitai:1971514@2231518",
23
+ "stats": {
24
+ "downloadCount": 47,
25
+ "ratingCount": 0,
26
+ "rating": 0,
27
+ "thumbsUpCount": 4
28
+ },
29
+ "model": {
30
+ "name": "Amia Lee (Gundam 00)",
31
+ "type": "LORA",
32
+ "nsfw": false,
33
+ "poi": false,
34
+ "description": "No Pony, No Illustrious, no any other, only 1.5",
35
+ "tags": [
36
+ "gundam",
37
+ "character"
38
+ ],
39
+ "allowNoCredit": true,
40
+ "allowCommercialUse": [
41
+ "Image",
42
+ "RentCivit",
43
+ "Rent",
44
+ "Sell"
45
+ ],
46
+ "allowDerivatives": true,
47
+ "allowDifferentLicense": true
48
+ },
49
+ "files": [
50
+ {
51
+ "id": 2124475,
52
+ "sizeKB": 18548.0859375,
53
+ "name": "AmiaLee-10.safetensors",
54
+ "type": "Model",
55
+ "pickleScanResult": "Success",
56
+ "pickleScanMessage": "No Pickle imports",
57
+ "virusScanResult": "Success",
58
+ "virusScanMessage": null,
59
+ "scannedAt": "2025-09-19T16:06:25.033Z",
60
+ "metadata": {
61
+ "format": "SafeTensor",
62
+ "size": null,
63
+ "fp": null
64
+ },
65
+ "hashes": {
66
+ "AutoV1": "DE2F2560",
67
+ "AutoV2": "BAD7B53BDD",
68
+ "SHA256": "BAD7B53BDDB69EA096DFDE3125B60D7D2B5222FA0C58A1448512864419AE15DB",
69
+ "CRC32": "BAB18BD4",
70
+ "BLAKE3": "BD78BE3BA87282EE9EE84CE04B99B57DBF95AC82871B4CA31908D758339444AD",
71
+ "AutoV3": "1499214C4B99"
72
+ },
73
+ "primary": true,
74
+ "downloadUrl": "https://civitai.com/api/download/models/2231518"
75
+ }
76
+ ],
77
+ "images": [
78
+ {
79
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/2f08a579-a8ce-46a5-82d7-5de885a81aa8/original=true/101252709.jpeg",
80
+ "nsfwLevel": 2,
81
+ "width": 512,
82
+ "height": 768,
83
+ "hash": "ULI}V40M~qR.~WNI?H%2?bR*xW%2%LxuWYR+",
84
+ "type": "image",
85
+ "metadata": {
86
+ "hash": "ULI}V40M~qR.~WNI?H%2?bR*xW%2%LxuWYR+",
87
+ "size": 194049,
88
+ "width": 512,
89
+ "height": 768
90
+ },
91
+ "minor": false,
92
+ "poi": false,
93
+ "meta": null,
94
+ "availability": "Public",
95
+ "hasMeta": false,
96
+ "hasPositivePrompt": false,
97
+ "onSite": false,
98
+ "remixOfId": null
99
+ },
100
+ {
101
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/47d32712-1d19-45a9-84fa-ffd380cd0eb1/original=true/101252711.jpeg",
102
+ "nsfwLevel": 2,
103
+ "width": 512,
104
+ "height": 768,
105
+ "hash": "UuN0@;t7~qxuM_afRjay%MofV@WCWBofayWB",
106
+ "type": "image",
107
+ "metadata": {
108
+ "hash": "UuN0@;t7~qxuM_afRjay%MofV@WCWBofayWB",
109
+ "size": 177738,
110
+ "width": 512,
111
+ "height": 768
112
+ },
113
+ "minor": false,
114
+ "poi": false,
115
+ "meta": null,
116
+ "availability": "Public",
117
+ "hasMeta": false,
118
+ "hasPositivePrompt": false,
119
+ "onSite": false,
120
+ "remixOfId": null
121
+ },
122
+ {
123
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/3c7177cd-2478-408d-8653-78d13125113d/original=true/101252710.jpeg",
124
+ "nsfwLevel": 2,
125
+ "width": 512,
126
+ "height": 768,
127
+ "hash": "UZKAyc_2~V?a~Vt7xut7xvkCj?R+tQazWXs:",
128
+ "type": "image",
129
+ "metadata": {
130
+ "hash": "UZKAyc_2~V?a~Vt7xut7xvkCj?R+tQazWXs:",
131
+ "size": 203814,
132
+ "width": 512,
133
+ "height": 768
134
+ },
135
+ "minor": false,
136
+ "poi": false,
137
+ "meta": null,
138
+ "availability": "Public",
139
+ "hasMeta": false,
140
+ "hasPositivePrompt": false,
141
+ "onSite": false,
142
+ "remixOfId": null
143
+ },
144
+ {
145
+ "url": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/45da7a10-e49c-45e6-9eb1-f0530f8c6d85/original=true/101252712.jpeg",
146
+ "nsfwLevel": 2,
147
+ "width": 512,
148
+ "height": 768,
149
+ "hash": "UZKwwoo#_3?b~pxutRxux]j]V@a}ofofNGax",
150
+ "type": "image",
151
+ "metadata": {
152
+ "hash": "UZKwwoo#_3?b~pxutRxux]j]V@a}ofofNGax",
153
+ "size": 214392,
154
+ "width": 512,
155
+ "height": 768
156
+ },
157
+ "minor": false,
158
+ "poi": false,
159
+ "meta": null,
160
+ "availability": "Public",
161
+ "hasMeta": false,
162
+ "hasPositivePrompt": false,
163
+ "onSite": false,
164
+ "remixOfId": null
165
+ }
166
+ ],
167
+ "downloadUrl": "https://civitai.com/api/download/models/2231518",
168
+ "creator": {
169
+ "username": "Shraffekyu",
170
+ "image": "https://image.civitai.com/xG1nkqKTMzGDvpLrqFT7WA/cf45bb38-3342-470e-9303-0e357944c367/width=96/Shraffekyu.jpeg"
171
+ },
172
+ "extensions": {
173
+ "sd_civitai_helper": {
174
+ "version": "1.8.13",
175
+ "last_update": 1760877727,
176
+ "skeleton_file": false
177
+ }
178
+ }
179
+ }
A-表番角色/高达系列/AmiaLee-10.preview.png ADDED

Git LFS Details

  • SHA256: 5740c70e130f627a79de58717745199e30b6f956261971dfcfcdc1ee664b6c82
  • Pointer size: 131 Bytes
  • Size of remote file: 194 kB
A-表番角色/高达系列/AmiaLee-10.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bad7b53bddb69ea096dfde3125b60d7d2b5222fa0c58a1448512864419ae15db
3
+ size 18993240