{ "last_node_id": 69, "last_link_id": 90, "nodes": [ { "id": 8, "type": "VAEDecode", "pos": { "0": 1083, "1": 19 }, "size": { "0": 210, "1": 46 }, "flags": {}, "order": 14, "mode": 0, "inputs": [ { "name": "samples", "type": "LATENT", "link": 24 }, { "name": "vae", "type": "VAE", "link": 12 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 60 ], "slot_index": 0 } ], "properties": { "Node name for S&R": "VAEDecode" } }, { "id": 22, "type": "BasicGuider", "pos": { "0": 783, "1": 212 }, "size": { "0": 241.79998779296875, "1": 46 }, "flags": {}, "order": 12, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 85, "slot_index": 0 }, { "name": "conditioning", "type": "CONDITIONING", "link": 74, "slot_index": 1 } ], "outputs": [ { "name": "GUIDER", "type": "GUIDER", "links": [ 30 ], "slot_index": 0, "shape": 3 } ], "properties": { "Node name for S&R": "BasicGuider" } }, { "id": 16, "type": "KSamplerSelect", "pos": { "0": 1110, "1": 648 }, "size": { "0": 315, "1": 58 }, "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [ { "name": "SAMPLER", "type": "SAMPLER", "links": [ 19 ], "shape": 3 } ], "properties": { "Node name for S&R": "KSamplerSelect" }, "widgets_values": [ "euler" ] }, { "id": 25, "type": "RandomNoise", "pos": { "0": 1128, "1": 353 }, "size": { "0": 315, "1": 82 }, "flags": {}, "order": 1, "mode": 0, "inputs": [], "outputs": [ { "name": "NOISE", "type": "NOISE", "links": [ 37 ], "shape": 3 } ], "properties": { "Node name for S&R": "RandomNoise" }, "widgets_values": [ 983245457248207, "randomize" ] }, { "id": 11, "type": "DualCLIPLoader", "pos": { "0": -473, "1": 356 }, "size": { "0": 315, "1": 106 }, "flags": {}, "order": 2, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP", "type": "CLIP", "links": [ 83 ], "slot_index": 0, "shape": 3 } ], "properties": { "Node name for S&R": "DualCLIPLoader" }, "widgets_values": [ "t5xxl_fp8_e4m3fn.safetensors", "clip_l.safetensors", "flux" ] }, { "id": 10, "type": "VAELoader", "pos": { "0": 611, "1": -42 }, "size": { "0": 315, "1": 58 }, "flags": {}, "order": 3, "mode": 0, "inputs": [], "outputs": [ { "name": "VAE", "type": "VAE", "links": [ 12, 64 ], "slot_index": 0, "shape": 3 } ], "properties": { "Node name for S&R": "VAELoader" }, "widgets_values": [ "ae.safetensors" ] }, { "id": 41, "type": "UpscaleModelLoader", "pos": { "0": 1188, "1": -183 }, "size": { "0": 315, "1": 58 }, "flags": {}, "order": 4, "mode": 0, "inputs": [], "outputs": [ { "name": "UPSCALE_MODEL", "type": "UPSCALE_MODEL", "links": [ 65 ], "shape": 3 } ], "properties": { "Node name for S&R": "UpscaleModelLoader" }, "widgets_values": [ "4x_foolhardy_Remacri.pth" ] }, { "id": 39, "type": "UltimateSDUpscale", "pos": { "0": 1462, "1": -59 }, "size": { "0": 315, "1": 614 }, "flags": {}, "order": 15, "mode": 4, "inputs": [ { "name": "image", "type": "IMAGE", "link": 60 }, { "name": "model", "type": "MODEL", "link": 61 }, { "name": "positive", "type": "CONDITIONING", "link": null, "slot_index": 2 }, { "name": "negative", "type": "CONDITIONING", "link": null, "slot_index": 3 }, { "name": "vae", "type": "VAE", "link": 64, "slot_index": 4 }, { "name": "upscale_model", "type": "UPSCALE_MODEL", "link": 65, "slot_index": 5 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "links": [ 66 ], "slot_index": 0, "shape": 3 } ], "properties": { "Node name for S&R": "UltimateSDUpscale" }, "widgets_values": [ 2, 439606151521400, "randomize", 20, 8, "euler", "simple", 0.2, "Linear", 512, 512, 8, 32, "None", 0, 64, 8, 16, true, false ] }, { "id": 66, "type": "Note", "pos": { "0": -145, "1": 971 }, "size": { "0": 445.3786315917969, "1": 251.9473419189453 }, "flags": {}, "order": 5, "mode": 0, "inputs": [], "outputs": [], "properties": { "text": "" }, "widgets_values": [ "young man, actor, intense expression, muddy clothes, t-shirt, jeans shorts, gym setting, dirt on body, hands clasped in front of body, brooding pose \n\nThe image shows a young man sitting on a wooden bench in a dark room. He is wearing a dark green t-shirt and grey pants. His hands are covered in mud and he appears to be deep in thought. His hair is messy and unkempt, and he has a serious expression on his face. The background is blurred, but it seems like he is in a workshop or garage. The overall mood of the image is somber and contemplative.\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 13, "type": "SamplerCustomAdvanced", "pos": { "0": 1099, "1": 171 }, "size": { "0": 355.20001220703125, "1": 106 }, "flags": {}, "order": 13, "mode": 0, "inputs": [ { "name": "noise", "type": "NOISE", "link": 37, "slot_index": 0 }, { "name": "guider", "type": "GUIDER", "link": 30, "slot_index": 1 }, { "name": "sampler", "type": "SAMPLER", "link": 19, "slot_index": 2 }, { "name": "sigmas", "type": "SIGMAS", "link": 20, "slot_index": 3 }, { "name": "latent_image", "type": "LATENT", "link": 89, "slot_index": 4 } ], "outputs": [ { "name": "output", "type": "LATENT", "links": [ 24 ], "slot_index": 0, "shape": 3 }, { "name": "denoised_output", "type": "LATENT", "links": null, "shape": 3 } ], "properties": { "Node name for S&R": "SamplerCustomAdvanced" } }, { "id": 68, "type": "Reroute", "pos": { "0": 1541.0958251953125, "1": 690.804443359375 }, "size": [ 75, 26 ], "flags": {}, "order": 8, "mode": 0, "inputs": [ { "name": "", "type": "*", "link": 90 } ], "outputs": [ { "name": "", "type": "LATENT", "links": [ 89 ] } ], "properties": { "showOutputText": false, "horizontal": false } }, { "id": 17, "type": "BasicScheduler", "pos": { "0": 1122, "1": 485 }, "size": { "0": 315, "1": 106 }, "flags": {}, "order": 9, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 38, "slot_index": 0 } ], "outputs": [ { "name": "SIGMAS", "type": "SIGMAS", "links": [ 20 ], "shape": 3 } ], "properties": { "Node name for S&R": "BasicScheduler" }, "widgets_values": [ "simple", 28, 1 ] }, { "id": 67, "type": "Empty Latent by Ratio (WLSH)", "pos": { "0": 1112, "1": 783 }, "size": { "0": 352.79998779296875, "1": 170 }, "flags": {}, "order": 6, "mode": 0, "inputs": [], "outputs": [ { "name": "latent", "type": "LATENT", "links": [ 90 ], "shape": 3 }, { "name": "width", "type": "INT", "links": null, "shape": 3 }, { "name": "height", "type": "INT", "links": null, "shape": 3 } ], "properties": { "Node name for S&R": "Empty Latent by Ratio (WLSH)" }, "widgets_values": [ "16:10", "portrait", 576, 2 ] }, { "id": 12, "type": "UNETLoader", "pos": { "0": -493, "1": 204 }, "size": { "0": 315, "1": 82 }, "flags": {}, "order": 7, "mode": 0, "inputs": [], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 38, 61, 82 ], "slot_index": 0, "shape": 3 } ], "properties": { "Node name for S&R": "UNETLoader" }, "widgets_values": [ "flux1-dev.safetensors", "default" ] }, { "id": 57, "type": "LoraLoader", "pos": { "0": -132, "1": 706 }, "size": { "0": 394.9391174316406, "1": 130.8064422607422 }, "flags": {}, "order": 10, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 82 }, { "name": "clip", "type": "CLIP", "link": 83 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 85 ], "slot_index": 0, "shape": 3 }, { "name": "CLIP", "type": "CLIP", "links": [ 87 ], "slot_index": 1, "shape": 3 } ], "properties": { "Node name for S&R": "LoraLoader" }, "widgets_values": [ "aodai_v2_000001250.safetensors", 0.8, 1 ] }, { "id": 43, "type": "CLIPTextEncodeFlux", "pos": { "0": 2126, "1": 576 }, "size": { "0": 432.8548278808594, "1": 348.077392578125 }, "flags": {}, "order": 11, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 87 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "links": [ 74 ], "slot_index": 0, "shape": 3 } ], "properties": { "Node name for S&R": "CLIPTextEncodeFlux" }, "widgets_values": [ "a photo of a young woman that is dressed in traditional Vietnamese dress called a0da1 with gold embroidery on the sleeves and neckline, looking to the left. She has long, straight, dark brown hair cascading down her back, she stands on a tiled floor in a grand, ornate room with classical columns and intricate carvings, suggesting an opulent or historical setting, The background features a large window allowing natural light to flood in, casting soft shadows and highlighting the contours of her body, the floor is tiled in a rich, reddish-brown color.", "a photo of a young woman that is dressed in traditional Vietnamese dress called a0da1 with gold embroidery on the sleeves and neckline, looking to the left. She has long, straight, dark brown hair cascading down her back, she stands on a tiled floor in a grand, ornate room with classical columns and intricate carvings, suggesting an opulent or historical setting, The background features a large window allowing natural light to flood in, casting soft shadows and highlighting the contours of her body, the floor is tiled in a rich, reddish-brown color.", 3.5 ], "color": "#232", "bgcolor": "#353" }, { "id": 29, "type": "Image Save", "pos": { "0": 1681, "1": 153 }, "size": { "0": 424.9637756347656, "1": 841.1275634765625 }, "flags": {}, "order": 16, "mode": 0, "inputs": [ { "name": "images", "type": "IMAGE", "link": 66 } ], "outputs": [ { "name": "images", "type": "IMAGE", "links": null, "shape": 3 }, { "name": "files", "type": "STRING", "links": null, "shape": 3 } ], "properties": { "Node name for S&R": "Image Save" }, "widgets_values": [ "flux[time(%Y-%m-%d)]", "ComfyUI", "_", 4, "false", "png", 300, 100, "true", "false", "false", "false", "true", "true", "true" ] } ], "links": [ [ 12, 10, 0, 8, 1, "VAE" ], [ 19, 16, 0, 13, 2, "SAMPLER" ], [ 20, 17, 0, 13, 3, "SIGMAS" ], [ 24, 13, 0, 8, 0, "LATENT" ], [ 30, 22, 0, 13, 1, "GUIDER" ], [ 37, 25, 0, 13, 0, "NOISE" ], [ 38, 12, 0, 17, 0, "MODEL" ], [ 60, 8, 0, 39, 0, "IMAGE" ], [ 61, 12, 0, 39, 1, "MODEL" ], [ 64, 10, 0, 39, 4, "VAE" ], [ 65, 41, 0, 39, 5, "UPSCALE_MODEL" ], [ 66, 39, 0, 29, 0, "IMAGE" ], [ 74, 43, 0, 22, 1, "CONDITIONING" ], [ 82, 12, 0, 57, 0, "MODEL" ], [ 83, 11, 0, 57, 1, "CLIP" ], [ 85, 57, 0, 22, 0, "MODEL" ], [ 87, 57, 1, 43, 0, "CLIP" ], [ 89, 68, 0, 13, 4, "LATENT" ], [ 90, 67, 0, 68, 0, "*" ] ], "groups": [], "config": {}, "extra": { "ds": { "scale": 1, "offset": [ -1099.5907249450684, -557.818227594549 ] } }, "version": 0.4 }