{ "id": "9a6bb580-900f-4a5f-9233-12eb0fe41226", "revision": 0, "last_node_id": 318, "last_link_id": 133, "nodes": [ { "id": 51, "type": "CLIPVisionEncode", "pos": [ 124.79998016357422, 510.9447021484375 ], "size": [ 253.60000610351562, 78 ], "flags": {}, "order": 21, "mode": 0, "inputs": [ { "name": "clip_vision", "type": "CLIP_VISION", "link": 94 }, { "name": "image", "type": "IMAGE", "link": 109 } ], "outputs": [ { "name": "CLIP_VISION_OUTPUT", "type": "CLIP_VISION_OUTPUT", "slot_index": 0, "links": [ 107 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPVisionEncode", "widget_ue_connectable": {} }, "widgets_values": [ "none" ] }, { "id": 8, "type": "VAEDecode", "pos": [ 1044.422607421875, 43.79334259033203 ], "size": [ 210, 46 ], "flags": { "collapsed": true }, "order": 32, "mode": 0, "inputs": [ { "name": "samples", "type": "LATENT", "link": 35 }, { "name": "vae", "type": "VAE", "link": 119 } ], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [ 112 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "VAEDecode", "widget_ue_connectable": {} }, "widgets_values": [] }, { "id": 54, "type": "ModelSamplingSD3", "pos": [ 602.1815185546875, 376.4954833984375 ], "size": [ 219.82144165039062, 59.586307525634766 ], "flags": {}, "order": 30, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 115 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "slot_index": 0, "links": [ 111 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "ModelSamplingSD3", "widget_ue_connectable": {} }, "widgets_values": [ 1.0000000000000002 ] }, { "id": 111, "type": "ModelPatchTorchSettings", "pos": [ 511.2032470703125, -189.2046661376953 ], "size": [ 253.60000610351562, 58 ], "flags": {}, "order": 29, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 114 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 115 ] } ], "properties": { "cnr_id": "comfyui-kjnodes", "ver": "5dcda71011870278c35d92ff77a677ed2e538f2d", "Node name for S&R": "ModelPatchTorchSettings", "widget_ue_connectable": {} }, "widgets_values": [ true ], "color": "#2a363b", "bgcolor": "#3f5159" }, { "id": 49, "type": "CLIPVisionLoader", "pos": [ 120.06415557861328, -152.35841369628906 ], "size": [ 315, 58 ], "flags": {}, "order": 0, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP_VISION", "type": "CLIP_VISION", "slot_index": 0, "links": [ 94 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPVisionLoader", "models": [ { "name": "clip_vision_h.safetensors", "url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/clip_vision/clip_vision_h.safetensors?download=true", "directory": "clip_vision" } ], "widget_ue_connectable": {} }, "widgets_values": [ "clip_vision_h.safetensors" ], "color": "#223", "bgcolor": "#335" }, { "id": 108, "type": "PathchSageAttentionKJ", "pos": [ 510.80859375, -310.6372985839844 ], "size": [ 253.63636779785156, 58 ], "flags": {}, "order": 28, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 129 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 114 ] } ], "properties": { "cnr_id": "comfyui-kjnodes", "ver": "5dcda71011870278c35d92ff77a677ed2e538f2d", "Node name for S&R": "PathchSageAttentionKJ", "widget_ue_connectable": {} }, "widgets_values": [ "auto" ], "color": "#2a363b", "bgcolor": "#3f5159" }, { "id": 297, "type": "UnetLoaderGGUFDisTorchMultiGPU", "pos": [ -357.08233642578125, -1244.5355224609375 ], "size": [ 413.7545166015625, 156.38363647460938 ], "flags": {}, "order": 1, "mode": 4, "inputs": [], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [] } ], "properties": { "cnr_id": "comfyui-multigpu", "ver": "1.7.3", "Node name for S&R": "UnetLoaderGGUFDisTorchMultiGPU", "widget_ue_connectable": {} }, "widgets_values": [ "wanvace\\Wan2.1_T2V_14B_FusionX_VACE-Q3_K_S.gguf", "cuda:0", 0.1, false, "" ], "color": "#322", "bgcolor": "#533" }, { "id": 299, "type": "MarkdownNote", "pos": [ -374.5340576171875, -1029.4305419921875 ], "size": [ 429.1710205078125, 183.4434356689453 ], "flags": {}, "order": 2, "mode": 0, "inputs": [], "outputs": [], "title": "speed boost", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## Just un-bypass the above node, and connect the output to where the current Unet Loader (GGUF) is connected then bypass that one." ], "color": "#432", "bgcolor": "#653" }, { "id": 298, "type": "MarkdownNote", "pos": [ 96.43013000488281, -1262.2916259765625 ], "size": [ 1014.274658203125, 420.11279296875 ], "flags": {}, "order": 3, "mode": 0, "inputs": [], "outputs": [], "title": "Low vram", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## ⬅️ Increase `virtual_vram_gb` to Simulate More VRAM\n\nUse this node instead of the standard **Unet Loader (GGUF)** if you're getting **out-of-memory (OOM)** errors.\n\n> ⚠️ **Note:** Higher values = more memory offloading = slower performance.\n\n---\n\n## ✅ Recommended Settings for 12GB and lower GPU (If you OOM)\n\n| **Field** | **Recommended Value** | **Description** |\n|---------------------------|------------------------------|---------------------------------------------------------------------------------|\n| `device` | `cuda:0` | Keep this set to use the primary GPU. |\n| `virtual_vram_gb` | `2.0` to `4.0` | Adds \"virtual\" VRAM by offloading to CPU. Start with 2.0 and increase as needed.|\n| `use_other_vram` | `true` | Enables fallback to CPU or additional GPU memory if available. |\n| `expert_mode_allocations`| *(leave blank unless advanced)* | Use only if you want manual control over layer/device allocation. |\n\n---\n" ], "color": "#322", "bgcolor": "#533" }, { "id": 301, "type": "MarkdownNote", "pos": [ 67.71233367919922, -767.65771484375 ], "size": [ 1097.5220947265625, 125.47974395751953 ], "flags": {}, "order": 4, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## 📦 Main Model should NOT be a FusionX main model and instead the normal base Wan2.1 14B image to video model since your using the FusionX LoRa here.\n\n\n\n\n\n\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 304, "type": "MarkdownNote", "pos": [ 1349.3433837890625, 846.6979370117188 ], "size": [ 586.1148681640625, 92.16458892822266 ], "flags": {}, "order": 5, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "

👈Change Video size and length here

\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 306, "type": "MarkdownNote", "pos": [ -315.1062927246094, 774.4031372070312 ], "size": [ 388.7239685058594, 88 ], "flags": {}, "order": 6, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "

Upload ref image here👉

\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 3, "type": "KSampler", "pos": [ 562.811279296875, 37.77678298950195 ], "size": [ 315, 262 ], "flags": {}, "order": 31, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 111 }, { "name": "positive", "type": "CONDITIONING", "link": 101 }, { "name": "negative", "type": "CONDITIONING", "link": 102 }, { "name": "latent_image", "type": "LATENT", "link": 103 } ], "outputs": [ { "name": "LATENT", "type": "LATENT", "slot_index": 0, "links": [ 35 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "KSampler", "widget_ue_connectable": {} }, "widgets_values": [ 1059002272748486, "randomize", 8, 1, "euler", "beta", 1 ] }, { "id": 7, "type": "CLIPTextEncode", "pos": [ 105.51646423339844, 281.5559997558594 ], "size": [ 425.27801513671875, 180.6060791015625 ], "flags": {}, "order": 20, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 133 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [ 98 ] } ], "title": "CLIP Text Encode (Negative Prompt)", "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {} }, "widgets_values": [ "色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多" ], "color": "#322", "bgcolor": "#533" }, { "id": 292, "type": "MarkdownNote", "pos": [ 801.4248046875, -212.00308227539062 ], "size": [ 349.3192443847656, 133.38710021972656 ], "flags": {}, "order": 7, "mode": 0, "inputs": [], "outputs": [], "title": "speed boost", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## ⬅️ If you don't have triton/sagatt then bypass these two nodes \n\n---\n\n\n\n\n\n\n\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 316, "type": "MarkdownNote", "pos": [ 1383.651123046875, -1129.8187255859375 ], "size": [ 1081.0716552734375, 231.4886016845703 ], "flags": {}, "order": 8, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## 🧪 Ingredients\n\nYou can now adjust these to get the custom look you want. For example, the **MPS LoRA** is known to cause some issues with faces — you can either bypass this node or set its value to `0` to disable it. ✨\n\nThe last two LoRAs were created by me! 🛠️ They add some extra detail and enhancements — feel free to tweak or turn them off depending on your preferences. 🎨\n\nThe default settings below are what’s used in the merge and LoRAs. ⚙️ \nCausVid and AccVid is meant for speed so I would not go under 0.5. Increase and use lower steps if you want to experiment.\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 308, "type": "MarkdownNote", "pos": [ 1418.4239501953125, -796.6294555664062 ], "size": [ 460.77166748046875, 132.94393920898438 ], "flags": {}, "order": 9, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## CausVidV2\n \n🔗 [Download Link](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_CausVid_14B_T2V_lora_rank32_v2.safetensors)\n\n🗂️ Place in: `ComfyUI/models/loras`\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 310, "type": "MarkdownNote", "pos": [ 1418.746337890625, -542.6290283203125 ], "size": [ 460.77166748046875, 132.94393920898438 ], "flags": {}, "order": 10, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## MPS Rewards\n \n🔗 [Download Link](https://huggingface.co/alibaba-pai/Wan2.1-Fun-Reward-LoRAs/blob/main/Wan2.1-Fun-14B-InP-MPS.safetensors)\n\n🗂️ Place in: `ComfyUI/models/loras`\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 309, "type": "MarkdownNote", "pos": [ 1428.8258056640625, -241.30262756347656 ], "size": [ 460.77166748046875, 132.94393920898438 ], "flags": {}, "order": 11, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## AccVid\n \n🔗 [Download Link](https://huggingface.co/Kijai/WanVideo_comfy/blob/main/Wan21_AccVid_I2V_480P_14B_lora_rank32_fp16.safetensors)\n\n🗂️ Place in: `ComfyUI/models/loras`\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 311, "type": "MarkdownNote", "pos": [ 1446.6082763671875, 88.24536895751953 ], "size": [ 460.77166748046875, 132.94393920898438 ], "flags": {}, "order": 12, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## Realism Boost\n \n🔗 [Download Link](https://huggingface.co/vrgamedevgirl84/Wan14BT2VFusioniX/resolve/main/OtherLoRa's/Wan14B_RealismBoost.safetensors?download=true)\n\n🗂️ Place in: `ComfyUI/models/loras`\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 315, "type": "MarkdownNote", "pos": [ 1450.012451171875, 404.9901123046875 ], "size": [ 460.77166748046875, 132.94393920898438 ], "flags": {}, "order": 13, "mode": 0, "inputs": [], "outputs": [], "title": "FusionX Lora", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## Detail Enhancer\n \n🔗 [Download Link](https://huggingface.co/vrgamedevgirl84/Wan14BT2VFusioniX/resolve/main/OtherLoRa's/DetailEnhancerV1.safetensors?download=true)\n\n🗂️ Place in: `ComfyUI/models/loras`\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 314, "type": "LoraLoaderModelOnly", "pos": [ 1911.3067626953125, -765.8120727539062 ], "size": [ 471.7080078125, 82 ], "flags": {}, "order": 22, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 128 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 123 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoraLoaderModelOnly", "widget_ue_connectable": {} }, "widgets_values": [ "WAN/Wan21_CausVid_14B_T2V_lora_rank32_v2.safetensors", 1.0000000000000002 ] }, { "id": 312, "type": "LoraLoaderModelOnly", "pos": [ 1925.962646484375, -518.8569946289062 ], "size": [ 470.5331726074219, 82 ], "flags": {}, "order": 24, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 123 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 124 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoraLoaderModelOnly", "widget_ue_connectable": {} }, "widgets_values": [ "WAN/Wan2.1-Fun-14B-InP-MPS.safetensors", 0.5000000000000001 ] }, { "id": 1, "type": "LoraLoaderModelOnly", "pos": [ 1958.6353759765625, -213.48617553710938 ], "size": [ 453.57525634765625, 82 ], "flags": {}, "order": 25, "mode": 0, "inputs": [ { "name": "model", "type": "MODEL", "link": 124 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 125 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoraLoaderModelOnly", "widget_ue_connectable": {} }, "widgets_values": [ "WAN/Wan21_AccVid_I2V_480P_14B_lora_rank32_fp16.safetensors", 1.0000000000000002 ] }, { "id": 318, "type": "CLIPLoaderGGUF", "pos": [ 141.15463256835938, -385.3298034667969 ], "size": [ 270, 82 ], "flags": {}, "order": 14, "mode": 0, "inputs": [], "outputs": [ { "name": "CLIP", "type": "CLIP", "links": [ 132, 133 ] } ], "properties": { "cnr_id": "comfyui-gguf", "ver": "1.1.0", "Node name for S&R": "CLIPLoaderGGUF" }, "widgets_values": [ "umt5-xxl-encoder-Q8_0.gguf", "wan" ] }, { "id": 288, "type": "MarkdownNote", "pos": [ -397.84307861328125, -577.1734619140625 ], "size": [ 457.41790771484375, 714.2341918945312 ], "flags": {}, "order": 15, "mode": 0, "inputs": [], "outputs": [], "title": "WAN - START HERE", "properties": { "widget_ue_connectable": {} }, "widgets_values": [ "## 📦 WAN Models to Download\n\n---\n\n### 🔴 Main WAN Model \n`Choose gguf from list' \n🔗 [Download Link](https://huggingface.co/city96/Wan2.1-I2V-14B-720P-gguf/tree/main)\n\n🗂️ Place in: `ComfyUI/models/diffusion_models`\n\n---\n\n### 🟣 WAN VAE \n`wan_2.1_vae.safetensors` \n🔗 [Download Link](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/blob/main/split_files/vae/wan_2.1_vae.safetensors)\n\n🗂️ Place in: `ComfyUI/models/vae`\n\n---\n### 🟣 WAN Text Encoder\n`GGUF text encoders'\n\n🔗 [Download Link](https://huggingface.co/city96/umt5-xxl-encoder-gguf/tree/main)\n\n🗂️ Place in: `ComfyUI/models/text_encoders`\n\n---\n### 🟣 WAN clip \n`clip_vision_h.safetensors.safetensors` \n🔗 [Download Link](https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/clip_vision/clip_vision_h.safetensors)\n\n🗂️ Place in: `ComfyUI/models/clip`\n\n\n\n\n\n\n" ], "color": "#432", "bgcolor": "#653" }, { "id": 52, "type": "LoadImage", "pos": [ 102.48811340332031, 639.4406127929688 ], "size": [ 315, 314 ], "flags": {}, "order": 16, "mode": 0, "inputs": [], "outputs": [ { "name": "IMAGE", "type": "IMAGE", "slot_index": 0, "links": [ 106, 109 ] }, { "name": "MASK", "type": "MASK", "slot_index": 1, "links": null } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoadImage", "widget_ue_connectable": {} }, "widgets_values": [ "vlcsnap-2025-06-27-17h35m18s519.png", "image" ] }, { "id": 305, "type": "VAELoader", "pos": [ 117.90325164794922, -265.6594543457031 ], "size": [ 320.67816162109375, 66.47000122070312 ], "flags": {}, "order": 17, "mode": 0, "inputs": [], "outputs": [ { "name": "VAE", "type": "VAE", "slot_index": 0, "links": [ 119, 120 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "VAELoader", "models": [ { "name": "wan_2.1_vae.safetensors", "url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true", "directory": "vae" } ], "widget_ue_connectable": {} }, "widgets_values": [ "Wan2.1_VAE.safetensors" ], "color": "#223", "bgcolor": "#335" }, { "id": 30, "type": "VHS_VideoCombine", "pos": [ 903.4950561523438, 120.0564956665039 ], "size": [ 423.81915283203125, 751.8191528320312 ], "flags": {}, "order": 33, "mode": 0, "inputs": [ { "name": "images", "type": "IMAGE", "link": 112 }, { "name": "audio", "shape": 7, "type": "AUDIO", "link": null }, { "name": "meta_batch", "shape": 7, "type": "VHS_BatchManager", "link": null }, { "name": "vae", "shape": 7, "type": "VAE", "link": null } ], "outputs": [ { "name": "Filenames", "type": "VHS_FILENAMES", "links": null } ], "properties": { "cnr_id": "comfyui-videohelpersuite", "ver": "1.6.1", "Node name for S&R": "VHS_VideoCombine", "widget_ue_connectable": {} }, "widgets_values": { "frame_rate": 16, "loop_count": 0, "filename_prefix": "FusionXi2v/FusionX", "format": "video/h264-mp4", "pix_fmt": "yuv420p", "crf": 19, "save_metadata": true, "trim_to_audio": false, "pingpong": false, "save_output": true, "videopreview": { "hidden": false, "paused": false, "params": { "filename": "FusionX_00023.mp4", "subfolder": "FusionXi2v", "type": "output", "format": "video/h264-mp4", "frame_rate": 16, "workflow": "FusionX_00023.png", "fullpath": "/home/berat/repositories/ComfyUI/output/FusionXi2v/FusionX_00023.mp4" } } } }, { "id": 295, "type": "UnetLoaderGGUF", "pos": [ 110.32615661621094, -513.88916015625 ], "size": [ 338.76776123046875, 72.42976379394531 ], "flags": {}, "order": 18, "mode": 0, "inputs": [], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 128 ] } ], "properties": { "cnr_id": "ComfyUI-GGUF", "ver": "6570efec6992015085f11b84e42d32f6cc71e8b7", "Node name for S&R": "UnetLoaderGGUF", "widget_ue_connectable": {} }, "widgets_values": [ "wan2.1-i2v-14b-480p-Q6_K.gguf" ], "color": "#223", "bgcolor": "#335" }, { "id": 50, "type": "WanImageToVideo", "pos": [ 973.5465698242188, 740.7857666015625 ], "size": [ 342.5999755859375, 210 ], "flags": {}, "order": 23, "mode": 0, "inputs": [ { "name": "positive", "type": "CONDITIONING", "link": 97 }, { "name": "negative", "type": "CONDITIONING", "link": 98 }, { "name": "vae", "type": "VAE", "link": 120 }, { "name": "clip_vision_output", "shape": 7, "type": "CLIP_VISION_OUTPUT", "link": 107 }, { "name": "start_image", "shape": 7, "type": "IMAGE", "link": 106 } ], "outputs": [ { "name": "positive", "type": "CONDITIONING", "slot_index": 0, "links": [ 101 ] }, { "name": "negative", "type": "CONDITIONING", "slot_index": 1, "links": [ 102 ] }, { "name": "latent", "type": "LATENT", "slot_index": 2, "links": [ 103 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "WanImageToVideo", "widget_ue_connectable": {} }, "widgets_values": [ 576, 576, 81, 1 ] }, { "id": 313, "type": "LoraLoaderModelOnly", "pos": [ 1991.4390869140625, 121.21511840820312 ], "size": [ 436.9215393066406, 87.3239974975586 ], "flags": {}, "order": 26, "mode": 4, "inputs": [ { "name": "model", "type": "MODEL", "link": 125 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 126 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoraLoaderModelOnly", "widget_ue_connectable": {} }, "widgets_values": [ "WAN/Wan14B_RealismBoost.safetensors", 0.4000000000000001 ] }, { "id": 4, "type": "LoraLoaderModelOnly", "pos": [ 2003.621337890625, 432.2743835449219 ], "size": [ 421.96795654296875, 82 ], "flags": {}, "order": 27, "mode": 4, "inputs": [ { "name": "model", "type": "MODEL", "link": 126 } ], "outputs": [ { "name": "MODEL", "type": "MODEL", "links": [ 129 ] } ], "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "LoraLoaderModelOnly", "widget_ue_connectable": {} }, "widgets_values": [ "WAN/DetailEnhancerV1.safetensors", 0.4000000000000001 ] }, { "id": 6, "type": "CLIPTextEncode", "pos": [ 111.7578353881836, 22.36838150024414 ], "size": [ 415.14501953125, 207.21304321289062 ], "flags": {}, "order": 19, "mode": 0, "inputs": [ { "name": "clip", "type": "CLIP", "link": 132 } ], "outputs": [ { "name": "CONDITIONING", "type": "CONDITIONING", "slot_index": 0, "links": [ 97 ] } ], "title": "CLIP Text Encode (Positive Prompt)", "properties": { "cnr_id": "comfy-core", "ver": "0.3.36", "Node name for S&R": "CLIPTextEncode", "widget_ue_connectable": {} }, "widgets_values": [ "Blue animal smooth walking animation like a lion. Animal is silent. Mouth is half closed." ], "color": "#232", "bgcolor": "#353" } ], "links": [ [ 35, 3, 0, 8, 0, "LATENT" ], [ 94, 49, 0, 51, 0, "CLIP_VISION" ], [ 97, 6, 0, 50, 0, "CONDITIONING" ], [ 98, 7, 0, 50, 1, "CONDITIONING" ], [ 101, 50, 0, 3, 1, "CONDITIONING" ], [ 102, 50, 1, 3, 2, "CONDITIONING" ], [ 103, 50, 2, 3, 3, "LATENT" ], [ 106, 52, 0, 50, 4, "IMAGE" ], [ 107, 51, 0, 50, 3, "CLIP_VISION_OUTPUT" ], [ 109, 52, 0, 51, 1, "IMAGE" ], [ 111, 54, 0, 3, 0, "MODEL" ], [ 112, 8, 0, 30, 0, "IMAGE" ], [ 114, 108, 0, 111, 0, "MODEL" ], [ 115, 111, 0, 54, 0, "MODEL" ], [ 119, 305, 0, 8, 1, "VAE" ], [ 120, 305, 0, 50, 2, "VAE" ], [ 123, 314, 0, 312, 0, "MODEL" ], [ 124, 312, 0, 1, 0, "MODEL" ], [ 125, 1, 0, 313, 0, "MODEL" ], [ 126, 313, 0, 4, 0, "MODEL" ], [ 128, 295, 0, 314, 0, "MODEL" ], [ 129, 4, 0, 108, 0, "MODEL" ], [ 132, 318, 0, 6, 0, "CLIP" ], [ 133, 318, 0, 7, 0, "CLIP" ] ], "groups": [ { "id": 2, "title": "IMAGE TO VIDEO", "bounding": [ 92.48811340332031, -61.26664733886719, 1244.8260498046875, 1024.707275390625 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 4, "title": "If you have triton/sagattn", "bounding": [ 496.4034423828125, -391.3373107910156, 284.74334716796875, 313.3325500488281 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 6, "title": "Group", "bounding": [ 100.32621765136719, -587.4893188476562, 370.55767822265625, 514.3302612304688 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 7, "title": "Group", "bounding": [ -384.5340881347656, -1335.8916015625, 1505.23876953125, 503.7127990722656 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 8, "title": "causvid", "bounding": [ 1408.4239501953125, -870.2294311523438, 984.589599609375, 216.54391479492188 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 9, "title": "mps", "bounding": [ 1408.746337890625, -616.2291870117188, 997.7484130859375, 216.5439453125 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 10, "title": "accvid", "bounding": [ 1418.8258056640625, -314.902587890625, 1003.3837890625, 216.5439453125 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 11, "title": "realismboost", "bounding": [ 1436.6082763671875, 14.645309448242188, 1001.751220703125, 216.54393005371094 ], "color": "#3f789e", "font_size": 24, "flags": {} }, { "id": 12, "title": "detail enhancer", "bounding": [ 1440.012451171875, 331.3905029296875, 995.5758056640625, 216.5439453125 ], "color": "#3f789e", "font_size": 24, "flags": {} } ], "config": {}, "extra": { "ds": { "scale": 0.6588450000000002, "offset": [ 254.98531330910748, 1071.7037238844844 ] }, "frontendVersion": "1.23.4", "ue_links": [], "links_added_by_ue": [], "VHS_latentpreview": false, "VHS_latentpreviewrate": 0, "VHS_MetadataImage": true, "VHS_KeepIntermediate": true }, "version": 0.4 }