Files
jaidaken f09734b0ee
Some checks failed
Python Linting / Run Ruff (push) Has been cancelled
Python Linting / Run Pylint (push) Has been cancelled
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.10, [self-hosted Linux], stable) (push) Has been cancelled
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.11, [self-hosted Linux], stable) (push) Has been cancelled
Full Comfy CI Workflow Runs / test-stable (12.1, , linux, 3.12, [self-hosted Linux], stable) (push) Has been cancelled
Full Comfy CI Workflow Runs / test-unix-nightly (12.1, , linux, 3.11, [self-hosted Linux], nightly) (push) Has been cancelled
Execution Tests / test (macos-latest) (push) Has been cancelled
Execution Tests / test (ubuntu-latest) (push) Has been cancelled
Execution Tests / test (windows-latest) (push) Has been cancelled
Test server launches without errors / test (push) Has been cancelled
Unit Tests / test (macos-latest) (push) Has been cancelled
Unit Tests / test (ubuntu-latest) (push) Has been cancelled
Unit Tests / test (windows-2022) (push) Has been cancelled
Add custom nodes, Civitai loras (LFS), and vast.ai setup script
Includes 30 custom nodes committed directly, 7 Civitai-exclusive
loras stored via Git LFS, and a setup script that installs all
dependencies and downloads HuggingFace-hosted models on vast.ai.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
2026-02-09 00:56:42 +00:00

625 lines
9.5 KiB
JSON

{
"id": "319b510b-b5ec-46d6-8605-a6a5fd7d6c6c",
"revision": 0,
"last_node_id": 25,
"last_link_id": 54,
"nodes": [
{
"id": 3,
"type": "KSampler",
"pos": [
1100,
620
],
"size": [
210,
474
],
"flags": {},
"order": 6,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 41
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 35
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 36
},
{
"name": "latent_image",
"type": "LATENT",
"link": 2
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
7
]
}
],
"properties": {
"Node name for S&R": "KSampler"
},
"widgets_values": [
0,
"fixed",
25,
7,
"euler",
"sgm_uniform",
1
]
},
{
"id": 5,
"type": "EmptyLatentImage",
"pos": [
580,
810
],
"size": [
210,
106
],
"flags": {},
"order": 0,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
2,
47
]
}
],
"properties": {
"Node name for S&R": "EmptyLatentImage"
},
"widgets_values": [
1024,
1024,
1
]
},
{
"id": 8,
"type": "VAEDecode",
"pos": [
1320,
620
],
"size": [
140,
46
],
"flags": {
"collapsed": false
},
"order": 8,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 7
},
{
"name": "vae",
"type": "VAE",
"link": 51
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
12
]
}
],
"properties": {
"Node name for S&R": "VAEDecode"
},
"widgets_values": []
},
{
"id": 13,
"type": "PreviewImage",
"pos": [
1320,
700
],
"size": [
440,
480
],
"flags": {},
"order": 10,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 12
}
],
"outputs": [],
"properties": {
"Node name for S&R": "PreviewImage"
},
"widgets_values": []
},
{
"id": 18,
"type": "NormalizedAttentionGuidance",
"pos": [
850,
620
],
"size": [
233.67147827148438,
198
],
"flags": {},
"order": 4,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 53
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 40
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
41
]
}
],
"properties": {
"Node name for S&R": "NormalizedAttentionGuidance"
},
"widgets_values": [
4,
0.5,
1,
-1,
10.000000000000002,
""
]
},
{
"id": 19,
"type": "CLIPTextEncode",
"pos": [
400,
530
],
"size": [
390,
100
],
"flags": {},
"order": 2,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 49
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
35,
45
]
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"elsa \\(frozen\\), portrait,"
]
},
{
"id": 20,
"type": "CLIPTextEncode",
"pos": [
400,
670
],
"size": [
390,
100
],
"flags": {},
"order": 3,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 50
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"links": [
36,
40,
46
]
}
],
"properties": {
"Node name for S&R": "CLIPTextEncode"
},
"widgets_values": [
"ugly, sketch, blurry, collage, blonde hair, blue eyes,"
]
},
{
"id": 22,
"type": "KSampler",
"pos": [
1100,
10
],
"size": [
210,
474
],
"flags": {},
"order": 5,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 54
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 45
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 46
},
{
"name": "latent_image",
"type": "LATENT",
"link": 47
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
43
]
}
],
"properties": {
"Node name for S&R": "KSampler"
},
"widgets_values": [
0,
"fixed",
25,
7,
"euler",
"sgm_uniform",
1
]
},
{
"id": 23,
"type": "PreviewImage",
"pos": [
1320,
90
],
"size": [
440,
480
],
"flags": {},
"order": 9,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 42
}
],
"outputs": [],
"properties": {
"Node name for S&R": "PreviewImage"
},
"widgets_values": []
},
{
"id": 24,
"type": "VAEDecode",
"pos": [
1320,
10
],
"size": [
140,
46
],
"flags": {
"collapsed": false
},
"order": 7,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 43
},
{
"name": "vae",
"type": "VAE",
"link": 52
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
42
]
}
],
"properties": {
"Node name for S&R": "VAEDecode"
},
"widgets_values": []
},
{
"id": 25,
"type": "CheckpointLoaderSimple",
"pos": [
400,
390
],
"size": [
390,
98
],
"flags": {},
"order": 1,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
53,
54
]
},
{
"name": "CLIP",
"type": "CLIP",
"links": [
49,
50
]
},
{
"name": "VAE",
"type": "VAE",
"links": [
51,
52
]
}
],
"properties": {
"Node name for S&R": "CheckpointLoaderSimple"
},
"widgets_values": [
"sdxl\\base\\sd_xl_base_1.0.safetensors"
]
}
],
"links": [
[
2,
5,
0,
3,
3,
"LATENT"
],
[
7,
3,
0,
8,
0,
"LATENT"
],
[
12,
8,
0,
13,
0,
"IMAGE"
],
[
35,
19,
0,
3,
1,
"CONDITIONING"
],
[
36,
20,
0,
3,
2,
"CONDITIONING"
],
[
40,
20,
0,
18,
1,
"CONDITIONING"
],
[
41,
18,
0,
3,
0,
"MODEL"
],
[
42,
24,
0,
23,
0,
"IMAGE"
],
[
43,
22,
0,
24,
0,
"LATENT"
],
[
45,
19,
0,
22,
1,
"CONDITIONING"
],
[
46,
20,
0,
22,
2,
"CONDITIONING"
],
[
47,
5,
0,
22,
3,
"LATENT"
],
[
49,
25,
1,
19,
0,
"CLIP"
],
[
50,
25,
1,
20,
0,
"CLIP"
],
[
51,
25,
2,
8,
1,
"VAE"
],
[
52,
25,
2,
24,
1,
"VAE"
],
[
53,
25,
0,
18,
0,
"MODEL"
],
[
54,
25,
0,
22,
0,
"MODEL"
]
],
"groups": [],
"config": {},
"extra": {
"frontendVersion": "1.23.0"
},
"version": 0.4
}