|
{ |
|
"last_node_id": 31, |
|
"last_link_id": 35, |
|
"nodes": [ |
|
{ |
|
"id": 1, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
395, |
|
222 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 0, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
1 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"PVCStyleModelMovable_pony160.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 2, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
391, |
|
397 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 1, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
2 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"accretiondiscxl_v10.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 3, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
823, |
|
284 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 15, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 1 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 2 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
4 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.71, |
|
0.67, |
|
0.75 |
|
] |
|
}, |
|
{ |
|
"id": 4, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
1208, |
|
284 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 16, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 3 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 4 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
5 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.58, |
|
0.5, |
|
0.6 |
|
] |
|
}, |
|
{ |
|
"id": 6, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
1544, |
|
519 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 2, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
6 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"bambooShootMixMix_v10.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 7, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
1613, |
|
280 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 17, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 5 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 6 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
7 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.5, |
|
0.4, |
|
0.5 |
|
] |
|
}, |
|
{ |
|
"id": 9, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
1969, |
|
544 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 3, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
8 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"cosmixv2WaifuToWife_v10.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 8, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
2035, |
|
281 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 18, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 7 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 8 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
9 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.44, |
|
0.33, |
|
0.42 |
|
] |
|
}, |
|
{ |
|
"id": 11, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
2410, |
|
526 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 4, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
10 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"malaSmoothPonyxl_v20.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 10, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
2423, |
|
277 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 19, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 9 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 10 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
11 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.4, |
|
0.25, |
|
0.37 |
|
] |
|
}, |
|
{ |
|
"id": 12, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
2773, |
|
534 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 5, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
12 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"meichidarkmixReload_meichidarkmixSensual.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 13, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
2869, |
|
267 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 20, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 11 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 12 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
14 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.37, |
|
0.25, |
|
0.33 |
|
] |
|
}, |
|
{ |
|
"id": 14, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
3243, |
|
521 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 6, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
13 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"mistoonXLCopper_v20Fast.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 17, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
3647, |
|
522 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 7, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
15 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"mocasemix_v10.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 15, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
3286, |
|
267 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 21, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 13 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 14 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
16 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.35000000000000003, |
|
0.22, |
|
0.3 |
|
] |
|
}, |
|
{ |
|
"id": 16, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
3673, |
|
263 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 22, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 15 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 16 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
18 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.33, |
|
0.2, |
|
0.27 |
|
] |
|
}, |
|
{ |
|
"id": 18, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
4015, |
|
518 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 8, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
17, |
|
21 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"alphonseWhiteDatura_Pony.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 19, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
4078, |
|
260 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 23, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 17 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 18 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
22 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.31, |
|
0.18, |
|
0.25 |
|
] |
|
}, |
|
{ |
|
"id": 20, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
4479, |
|
264 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 24, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 21 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 22 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
23 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.3, |
|
0.16, |
|
0.23 |
|
] |
|
}, |
|
{ |
|
"id": 22, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
4475, |
|
522 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 9, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
24 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"pixivponyepitamix_v10.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 24, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
4889, |
|
500 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 10, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
25 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"pixivponyepitamix_v10_1.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 21, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
4890, |
|
281 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 25, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 24 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 23 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
26 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.28, |
|
0.15, |
|
0.21 |
|
] |
|
}, |
|
{ |
|
"id": 23, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
5265, |
|
282 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 26, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 25 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 26 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
27 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.27, |
|
0.14, |
|
0.2 |
|
] |
|
}, |
|
{ |
|
"id": 25, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
5290, |
|
485 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 11, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
28 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"thickCoatingStyle_pdxl10.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 26, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
5680, |
|
260 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 27, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 28 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 27 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
29 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.26, |
|
0.13, |
|
0.18 |
|
] |
|
}, |
|
{ |
|
"id": 27, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
5683, |
|
456 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 12, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
30 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts." |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"vendoPonyRealistic_v13Lora.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 28, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
6063, |
|
249 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 28, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 30 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 29 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
31 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.25, |
|
0.12, |
|
0.17 |
|
] |
|
}, |
|
{ |
|
"id": 31, |
|
"type": "CheckpointSave", |
|
"pos": [ |
|
6887, |
|
258 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 30, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model", |
|
"type": "MODEL", |
|
"link": 33 |
|
}, |
|
{ |
|
"name": "clip", |
|
"type": "CLIP", |
|
"link": 34 |
|
}, |
|
{ |
|
"name": "vae", |
|
"type": "VAE", |
|
"link": 35 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointSave" |
|
}, |
|
"widgets_values": [ |
|
"checkpoints/ComfyUI" |
|
] |
|
}, |
|
{ |
|
"id": 29, |
|
"type": "ModelMergeBlocks", |
|
"pos": [ |
|
6425, |
|
247 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 126 |
|
}, |
|
"flags": {}, |
|
"order": 29, |
|
"mode": 0, |
|
"inputs": [ |
|
{ |
|
"name": "model1", |
|
"type": "MODEL", |
|
"link": 32 |
|
}, |
|
{ |
|
"name": "model2", |
|
"type": "MODEL", |
|
"link": 31 |
|
} |
|
], |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
33 |
|
], |
|
"shape": 3, |
|
"slot_index": 0 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "ModelMergeBlocks" |
|
}, |
|
"widgets_values": [ |
|
0.25, |
|
0.11, |
|
0.16 |
|
] |
|
}, |
|
{ |
|
"id": 30, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
6048, |
|
452 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 13, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
32 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts.", |
|
"slot_index": 1 |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": null, |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space." |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"whiteUnicorn_v30.safetensors" |
|
] |
|
}, |
|
{ |
|
"id": 5, |
|
"type": "CheckpointLoaderSimple", |
|
"pos": [ |
|
1035, |
|
519 |
|
], |
|
"size": { |
|
"0": 315, |
|
"1": 98 |
|
}, |
|
"flags": {}, |
|
"order": 14, |
|
"mode": 0, |
|
"outputs": [ |
|
{ |
|
"name": "MODEL", |
|
"type": "MODEL", |
|
"links": [ |
|
3 |
|
], |
|
"shape": 3, |
|
"tooltip": "The model used for denoising latents.", |
|
"slot_index": 0 |
|
}, |
|
{ |
|
"name": "CLIP", |
|
"type": "CLIP", |
|
"links": [ |
|
34 |
|
], |
|
"shape": 3, |
|
"tooltip": "The CLIP model used for encoding text prompts.", |
|
"slot_index": 1 |
|
}, |
|
{ |
|
"name": "VAE", |
|
"type": "VAE", |
|
"links": [ |
|
35 |
|
], |
|
"shape": 3, |
|
"tooltip": "The VAE model used for encoding and decoding images to and from latent space.", |
|
"slot_index": 2 |
|
} |
|
], |
|
"properties": { |
|
"Node name for S&R": "CheckpointLoaderSimple" |
|
}, |
|
"widgets_values": [ |
|
"pixivponyepitamix_v10.safetensors" |
|
] |
|
} |
|
], |
|
"links": [ |
|
[ |
|
1, |
|
1, |
|
0, |
|
3, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
2, |
|
2, |
|
0, |
|
3, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
3, |
|
5, |
|
0, |
|
4, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
4, |
|
3, |
|
0, |
|
4, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
5, |
|
4, |
|
0, |
|
7, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
6, |
|
6, |
|
0, |
|
7, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
7, |
|
7, |
|
0, |
|
8, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
8, |
|
9, |
|
0, |
|
8, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
9, |
|
8, |
|
0, |
|
10, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
10, |
|
11, |
|
0, |
|
10, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
11, |
|
10, |
|
0, |
|
13, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
12, |
|
12, |
|
0, |
|
13, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
13, |
|
14, |
|
0, |
|
15, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
14, |
|
13, |
|
0, |
|
15, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
15, |
|
17, |
|
0, |
|
16, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
16, |
|
15, |
|
0, |
|
16, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
17, |
|
18, |
|
0, |
|
19, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
18, |
|
16, |
|
0, |
|
19, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
21, |
|
18, |
|
0, |
|
20, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
22, |
|
19, |
|
0, |
|
20, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
23, |
|
20, |
|
0, |
|
21, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
24, |
|
22, |
|
0, |
|
21, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
25, |
|
24, |
|
0, |
|
23, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
26, |
|
21, |
|
0, |
|
23, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
27, |
|
23, |
|
0, |
|
26, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
28, |
|
25, |
|
0, |
|
26, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
29, |
|
26, |
|
0, |
|
28, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
30, |
|
27, |
|
0, |
|
28, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
31, |
|
28, |
|
0, |
|
29, |
|
1, |
|
"MODEL" |
|
], |
|
[ |
|
32, |
|
30, |
|
0, |
|
29, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
33, |
|
29, |
|
0, |
|
31, |
|
0, |
|
"MODEL" |
|
], |
|
[ |
|
34, |
|
5, |
|
1, |
|
31, |
|
1, |
|
"CLIP" |
|
], |
|
[ |
|
35, |
|
5, |
|
2, |
|
31, |
|
2, |
|
"VAE" |
|
] |
|
], |
|
"groups": [], |
|
"config": {}, |
|
"extra": { |
|
"ds": { |
|
"scale": 1, |
|
"offset": [ |
|
-5635.498001420395, |
|
96.06443919089634 |
|
] |
|
} |
|
}, |
|
"version": 0.4 |
|
} |