Motion-Lora-Camera-Push-In-Wan-14B-720p-I2V / workflow-pushin-v2-comfyui.json
lovisdotio
main
e1ebfaf
{
"id": "9a6bb580-900f-4a5f-9233-12eb0fe41226",
"revision": 0,
"last_node_id": 347,
"last_link_id": 169,
"nodes": [
{
"id": 51,
"type": "CLIPVisionEncode",
"pos": [
199.71192932128906,
524.9036254882812
],
"size": [
253.60000610351562,
78
],
"flags": {},
"order": 20,
"mode": 0,
"inputs": [
{
"name": "clip_vision",
"type": "CLIP_VISION",
"link": 94
},
{
"name": "image",
"type": "IMAGE",
"link": 145
}
],
"outputs": [
{
"name": "CLIP_VISION_OUTPUT",
"type": "CLIP_VISION_OUTPUT",
"slot_index": 0,
"links": [
107
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "CLIPVisionEncode",
"widget_ue_connectable": {}
},
"widgets_values": [
"none"
]
},
{
"id": 39,
"type": "VAELoader",
"pos": [
546.8587646484375,
-291.62469482421875
],
"size": [
306.36004638671875,
58
],
"flags": {},
"order": 0,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "VAE",
"type": "VAE",
"slot_index": 0,
"links": [
76,
99
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "VAELoader",
"models": [
{
"name": "wan_2.1_vae.safetensors",
"url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/vae/wan_2.1_vae.safetensors?download=true",
"directory": "vae"
}
],
"widget_ue_connectable": {}
},
"widgets_values": [
"wan_2.1_vae.safetensors"
]
},
{
"id": 49,
"type": "CLIPVisionLoader",
"pos": [
543.3070678710938,
-166.9625244140625
],
"size": [
315,
58
],
"flags": {},
"order": 1,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CLIP_VISION",
"type": "CLIP_VISION",
"slot_index": 0,
"links": [
94
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "CLIPVisionLoader",
"models": [
{
"name": "clip_vision_h.safetensors",
"url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/clip_vision/clip_vision_h.safetensors?download=true",
"directory": "clip_vision"
}
],
"widget_ue_connectable": {}
},
"widgets_values": [
"clip_vision_h.safetensors"
]
},
{
"id": 7,
"type": "CLIPTextEncode",
"pos": [
18.697704315185547,
316.1113586425781
],
"size": [
594.2410888671875,
170.46165466308594
],
"flags": {},
"order": 15,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 75
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
98
]
}
],
"title": "CLIP Text Encode (Negative Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "CLIPTextEncode",
"widget_ue_connectable": {}
},
"widgets_values": [
"色调艳丽,过曝,静态,细节模糊不清,字幕,风格,作品,画作,画面,静止,整体发灰,最差质量,低质量,JPEG压缩残留,丑陋的,残缺的,多余的手指,画得不好的手部,画得不好的脸部,畸形的,毁容的,形态畸形的肢体,手指融合,静止不动的画面,杂乱的背景,三条腿,背景人很多,倒着走"
],
"color": "#322",
"bgcolor": "#533"
},
{
"id": 6,
"type": "CLIPTextEncode",
"pos": [
186.66978454589844,
36.32728958129883
],
"size": [
415.14501953125,
207.21304321289062
],
"flags": {},
"order": 25,
"mode": 0,
"inputs": [
{
"name": "clip",
"type": "CLIP",
"link": 74
},
{
"name": "text",
"type": "STRING",
"widget": {
"name": "text"
},
"link": 142
}
],
"outputs": [
{
"name": "CONDITIONING",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
97
]
}
],
"title": "CLIP Text Encode (Positive Prompt)",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "CLIPTextEncode",
"widget_ue_connectable": {}
},
"widgets_values": [
"beautiful forest"
],
"color": "#232",
"bgcolor": "#353"
},
{
"id": 38,
"type": "CLIPLoader",
"pos": [
126.74717712402344,
-193.13661193847656
],
"size": [
387.79998779296875,
106
],
"flags": {},
"order": 2,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "CLIP",
"type": "CLIP",
"slot_index": 0,
"links": [
74,
75
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "CLIPLoader",
"models": [
{
"name": "umt5_xxl_fp8_e4m3fn_scaled.safetensors",
"url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/text_encoders/umt5_xxl_fp8_e4m3fn_scaled.safetensors?download=true",
"directory": "text_encoders"
}
],
"widget_ue_connectable": {}
},
"widgets_values": [
"umt5_xxl_fp16.safetensors",
"wan",
"default"
]
},
{
"id": 331,
"type": "LoadUpscalerTensorrtModel",
"pos": [
2395.3759765625,
127.51609802246094
],
"size": [
281.9956970214844,
82
],
"flags": {},
"order": 3,
"mode": 4,
"inputs": [],
"outputs": [
{
"name": "upscaler_trt_model",
"type": "UPSCALER_TRT_MODEL",
"links": [
147
]
}
],
"properties": {
"cnr_id": "ComfyUI-Upscaler-Tensorrt",
"ver": "e0c94a8dac0310bb8d6add6838db4ed137106da8",
"Node name for S&R": "LoadUpscalerTensorrtModel"
},
"widgets_values": [
"RealESRGAN_x4",
"fp16"
]
},
{
"id": 332,
"type": "UpscalerTensorrt",
"pos": [
2750.006591796875,
103.88851928710938
],
"size": [
270,
78
],
"flags": {},
"order": 31,
"mode": 4,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 148
},
{
"name": "upscaler_trt_model",
"type": "UPSCALER_TRT_MODEL",
"link": 147
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
162
]
}
],
"properties": {
"cnr_id": "ComfyUI-Upscaler-Tensorrt",
"ver": "e0c94a8dac0310bb8d6add6838db4ed137106da8",
"Node name for S&R": "UpscalerTensorrt"
},
"widgets_values": [
"2k"
]
},
{
"id": 333,
"type": "VHS_VideoCombine",
"pos": [
3161.56396484375,
110.23788452148438
],
"size": [
423.81915283203125,
575.1482543945312
],
"flags": {},
"order": 32,
"mode": 4,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 162
},
{
"name": "audio",
"shape": 7,
"type": "AUDIO",
"link": null
},
{
"name": "meta_batch",
"shape": 7,
"type": "VHS_BatchManager",
"link": null
},
{
"name": "vae",
"shape": 7,
"type": "VAE",
"link": null
}
],
"outputs": [
{
"name": "Filenames",
"type": "VHS_FILENAMES",
"links": null
}
],
"properties": {
"cnr_id": "comfyui-videohelpersuite",
"ver": "1.6.1",
"Node name for S&R": "VHS_VideoCombine",
"widget_ue_connectable": {}
},
"widgets_values": {
"frame_rate": 25,
"loop_count": 0,
"filename_prefix": "GOODFLY2/GOOD",
"format": "video/h264-mp4",
"pix_fmt": "yuv420p",
"crf": 17,
"save_metadata": true,
"trim_to_audio": false,
"pingpong": false,
"save_output": true,
"videopreview": {
"hidden": false,
"paused": false,
"params": {
"filename": "GOOD_00001.mp4",
"subfolder": "GOODFLY2",
"type": "output",
"format": "video/h264-mp4",
"frame_rate": 25,
"workflow": "GOOD_00001.png",
"fullpath": "/workspace/ComfyUI/output/GOODFLY2/GOOD_00001.mp4"
}
}
}
},
{
"id": 325,
"type": "ttN concat",
"pos": [
-440.1773681640625,
310.7658386230469
],
"size": [
400,
220
],
"flags": {},
"order": 22,
"mode": 0,
"inputs": [
{
"name": "text2",
"type": "STRING",
"widget": {
"name": "text2"
},
"link": 141
}
],
"outputs": [
{
"name": "concat",
"type": "STRING",
"links": [
142
]
}
],
"properties": {
"cnr_id": "comfyui_tinyterranodes",
"ver": "2.0.7",
"Node name for S&R": "ttN concat",
"ttNnodeVersion": "1.0.0"
},
"widgets_values": [
"Push-in camera, ",
"",
"",
""
]
},
{
"id": 342,
"type": "Note",
"pos": [
159.68435668945312,
-716.724365234375
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 4,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"LoRa that allow to render only in 4 steps"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 343,
"type": "Note",
"pos": [
157.97186279296875,
-514.5469970703125
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 5,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"You can play with the motionpushin lora between value from 0.5 to 1.2"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 37,
"type": "UNETLoader",
"pos": [
119.10247802734375,
-315.3940734863281
],
"size": [
376.7245178222656,
82
],
"flags": {},
"order": 6,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"slot_index": 0,
"links": [
169
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "UNETLoader",
"models": [
{
"name": "wan2.1_i2v_480p_14B_fp16.safetensors",
"url": "https://huggingface.co/Comfy-Org/Wan_2.1_ComfyUI_repackaged/resolve/main/split_files/diffusion_models/wan2.1_i2v_480p_14B_fp16.safetensors?download=true",
"directory": "diffusion_models"
}
],
"widget_ue_connectable": {}
},
"widgets_values": [
"wan2.1_i2v_720p_14B_fp16.safetensors",
"default"
]
},
{
"id": 108,
"type": "PathchSageAttentionKJ",
"pos": [
1033.7701416015625,
-409.9872741699219
],
"size": [
270,
58
],
"flags": {},
"order": 24,
"mode": 4,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 165
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
114
]
}
],
"properties": {
"cnr_id": "comfyui-kjnodes",
"ver": "5dcda71011870278c35d92ff77a677ed2e538f2d",
"Node name for S&R": "PathchSageAttentionKJ",
"widget_ue_connectable": {}
},
"widgets_values": [
"auto"
]
},
{
"id": 111,
"type": "ModelPatchTorchSettings",
"pos": [
1042.7601318359375,
-289.77960205078125
],
"size": [
307.443359375,
58
],
"flags": {},
"order": 26,
"mode": 4,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 114
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": []
}
],
"properties": {
"cnr_id": "comfyui-kjnodes",
"ver": "5dcda71011870278c35d92ff77a677ed2e538f2d",
"Node name for S&R": "ModelPatchTorchSettings",
"widget_ue_connectable": {}
},
"widgets_values": [
true
]
},
{
"id": 344,
"type": "Note",
"pos": [
-247.32579040527344,
-334.1845397949219
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 7,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"the Lora Model \"motionpushin\" is train on Wan 2.1 I2V 720p 14B. It is possible that"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 52,
"type": "LoadImage",
"pos": [
-1702.1080322265625,
-85.75054931640625
],
"size": [
354.39892578125,
326
],
"flags": {},
"order": 8,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
139,
144
]
},
{
"name": "MASK",
"type": "MASK",
"slot_index": 1,
"links": null
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "LoadImage",
"widget_ue_connectable": {}
},
"widgets_values": [
"lovis.io_POV_darting_50_cm_above_a_Provence_lavender_field_at_f1b8fd89-2636-4696-ac52-609a361acfd7_1.png",
"image"
],
"color": "#323",
"bgcolor": "#535"
},
{
"id": 341,
"type": "Note",
"pos": [
-788.6988525390625,
298.69390869140625
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 9,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"\"Push-in camera\" is the trigger for the Lora Model"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 345,
"type": "Note",
"pos": [
-1702.7252197265625,
-315.1345520019531
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 10,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"The workflow is fully automatic as its using Llm, so if you just put an image as input, it should work"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 30,
"type": "VHS_VideoCombine",
"pos": [
1729.515625,
97.9158935546875
],
"size": [
418.4777526855469,
572.1437377929688
],
"flags": {},
"order": 30,
"mode": 0,
"inputs": [
{
"name": "images",
"type": "IMAGE",
"link": 112
},
{
"name": "audio",
"shape": 7,
"type": "AUDIO",
"link": null
},
{
"name": "meta_batch",
"shape": 7,
"type": "VHS_BatchManager",
"link": null
},
{
"name": "vae",
"shape": 7,
"type": "VAE",
"link": null
}
],
"outputs": [
{
"name": "Filenames",
"type": "VHS_FILENAMES",
"links": null
}
],
"properties": {
"cnr_id": "comfyui-videohelpersuite",
"ver": "1.6.1",
"Node name for S&R": "VHS_VideoCombine",
"widget_ue_connectable": {}
},
"widgets_values": {
"frame_rate": 25,
"loop_count": 0,
"filename_prefix": "GOODFLY3/GOOD",
"format": "video/h264-mp4",
"pix_fmt": "yuv420p10le",
"crf": 17,
"save_metadata": true,
"trim_to_audio": false,
"pingpong": false,
"save_output": true,
"videopreview": {
"hidden": false,
"paused": false,
"params": {
"filename": "GOOD_00091.mp4",
"subfolder": "GOODFLY3",
"type": "output",
"format": "video/h264-mp4",
"frame_rate": 25,
"workflow": "GOOD_00091.png",
"fullpath": "/workspace/ComfyUI/output/GOODFLY3/GOOD_00091.mp4"
}
}
},
"color": "#222",
"bgcolor": "#000"
},
{
"id": 8,
"type": "VAEDecode",
"pos": [
1459.3970947265625,
109.40514373779297
],
"size": [
210,
46
],
"flags": {
"collapsed": false
},
"order": 29,
"mode": 0,
"inputs": [
{
"name": "samples",
"type": "LATENT",
"link": 35
},
{
"name": "vae",
"type": "VAE",
"link": 76
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"slot_index": 0,
"links": [
112,
148
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "VAEDecode",
"widget_ue_connectable": {}
},
"widgets_values": []
},
{
"id": 3,
"type": "KSampler",
"pos": [
1116.6273193359375,
81.61148834228516
],
"size": [
315,
262
],
"flags": {},
"order": 28,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 159
},
{
"name": "positive",
"type": "CONDITIONING",
"link": 101
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 102
},
{
"name": "latent_image",
"type": "LATENT",
"link": 103
}
],
"outputs": [
{
"name": "LATENT",
"type": "LATENT",
"slot_index": 0,
"links": [
35
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "KSampler",
"widget_ue_connectable": {}
},
"widgets_values": [
481828910680033,
"randomize",
4,
2,
"euler",
"beta",
1
]
},
{
"id": 54,
"type": "ModelSamplingSD3",
"pos": [
848.4885864257812,
-44.23977279663086
],
"size": [
219.82144165039062,
59.586307525634766
],
"flags": {
"collapsed": false
},
"order": 23,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 164
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"slot_index": 0,
"links": [
159
]
}
],
"title": "Shift",
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "ModelSamplingSD3",
"widget_ue_connectable": {}
},
"widgets_values": [
8.000000000000002
]
},
{
"id": 330,
"type": "ImageScale",
"pos": [
727.0180053710938,
448.8490905761719
],
"size": [
270,
130
],
"flags": {},
"order": 17,
"mode": 0,
"inputs": [
{
"name": "image",
"type": "IMAGE",
"link": 144
}
],
"outputs": [
{
"name": "IMAGE",
"type": "IMAGE",
"links": [
145,
146
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.44",
"Node name for S&R": "ImageScale"
},
"widgets_values": [
"nearest-exact",
1280,
720,
"center"
]
},
{
"id": 50,
"type": "WanImageToVideo",
"pos": [
719.2560424804688,
161.1350860595703
],
"size": [
342.5999755859375,
210
],
"flags": {},
"order": 27,
"mode": 0,
"inputs": [
{
"name": "positive",
"type": "CONDITIONING",
"link": 97
},
{
"name": "negative",
"type": "CONDITIONING",
"link": 98
},
{
"name": "vae",
"type": "VAE",
"link": 99
},
{
"name": "clip_vision_output",
"shape": 7,
"type": "CLIP_VISION_OUTPUT",
"link": 107
},
{
"name": "start_image",
"shape": 7,
"type": "IMAGE",
"link": 146
}
],
"outputs": [
{
"name": "positive",
"type": "CONDITIONING",
"slot_index": 0,
"links": [
101
]
},
{
"name": "negative",
"type": "CONDITIONING",
"slot_index": 1,
"links": [
102
]
},
{
"name": "latent",
"type": "LATENT",
"slot_index": 2,
"links": [
103
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "WanImageToVideo",
"widget_ue_connectable": {}
},
"widgets_values": [
1280,
720,
81,
1
]
},
{
"id": 340,
"type": "Note",
"pos": [
-756.67333984375,
611.007080078125
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 11,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"Dont forget here to put your api_key for using Google AI service. Can be created here https://aistudio.google.com/\n\nI use the node of Fill node, normally its made for video, but its possible to only use one image. "
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 347,
"type": "Note",
"pos": [
-1223.114990234375,
817.5936279296875
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 12,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"This is the instruction for Gemini. change it at your convenience. "
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 328,
"type": "Text Multiline",
"pos": [
-842.7467651367188,
803.4328002929688
],
"size": [
387.0903625488281,
297.1717529296875
],
"flags": {},
"order": 13,
"mode": 0,
"inputs": [],
"outputs": [
{
"name": "STRING",
"type": "STRING",
"links": [
143
]
}
],
"properties": {
"cnr_id": "was-node-suite-comfyui",
"ver": "3ed45af34a14551dc28cb3127235cc7197d4633f",
"Node name for S&R": "Text Multiline"
},
"widgets_values": [
"Instructions for Gemini API: Generating Hyper-Detailed, Rapid Push-in Animation Prompts\nInput: An image.\n\nOutput: A single, exceptionally detailed and highly dynamic text prompt, meticulously describing a super-fast, aggressive \"push-in\" animation based on the input image.\n\nCore Instructions for Gemini:\nYour primary task is to act as an advanced animation prompt generator, creating prompts that are not only dynamic but also exceptionally rich in visual information, capturing the full essence of the input image. When you receive an image, you must follow these steps to create a highly descriptive and dynamic prompt:\n\nForensic Image Analysis (Maximum Detail):\n\nPerform an exhaustive analysis of the image, capturing every discernible visual element, texture, and nuance.\n\nIdentify and describe:\n\nMain subjects and their specific characteristics: What are they? What material are they made of? Are there any unique markings, patterns, or intricate designs?\n\nForeground, midground, and background elements: Detail what is present at each depth, noting their relationship to the overall scene.\n\nSpecific objects and their states: Are they worn, new, sharp, blurred?\n\nTextures: Describe the tactile quality of surfaces (e.g., rough granite, slick metallic, shimmering silk, cracked earth, wet leaves).\n\nLighting conditions: Go beyond just \"day\" or \"night.\" Describe the quality, direction, and color of light (e.g., harsh midday glare, soft golden hour luminescence, diffused overcast glow, sharp dramatic shadows, flickering artificial light, ethereal moonbeams). Note how light interacts with surfaces – reflections, refractions, glints.\n\nColors: Specify dominant hues, subtle color variations, and how colors interplay (e.g., muted earth tones, vibrant contrasting primaries, iridescent shifts).\n\nAtmosphere and mood: Is it misty, clear, dusty, serene, chaotic, melancholic, futuristic, ancient? Use evocative adjectives.\n\nEnvironmental specificities: If a natural scene, identify specific flora/fauna, geological features. If urban, architectural styles, types of vehicles, signs.\n\nMicro-details: Zoom in mentally. Are there tiny cracks, dust motes, water droplets, individual leaves, specific patterns on fabrics?\n\nElaborate on the \"Push-in\" Motion (Super Fast & Visually Integrated):\n\nThe core characteristic of the movement MUST be speed, intensity, and a sense of aggressive, almost overwhelming, progression.\n\nIntegrate the minute details from step 1 into the description of the movement, showing what the camera rapidly passes through, by, or into.\n\nIf the image depicts a traversable landscape or scene:\n\nDescribe the camera rapidly plunging, rocketing, or violently streaking through the environment.\n\nEmphasize the blurring and rapid distortion of passing elements due to high velocity, and the sudden shift in perspective as details rush into focus.\n\nExamples: \"Violently streaking through an ancient, gnarled forest, the vibrant moss-covered roots and individual, dew-kissed fern fronds blurring into streaks as blinding shafts of sunlight aggressively pierce the incredibly dense, emerald canopy, revealing a dimly lit, winding path of damp, dark earth that vanishes into the depths.\"\n\nIf the image depicts a static or non-traversable scene:\n\nDescribe a lightning-fast, aggressive lunge or instantaneous acceleration towards a specific, highly detailed focal point.\n\nEmphasize the abruptness of the reveal, where intricate, previously unseen details burst into sharp clarity as the camera slams into position.\n\nExamples: \"Instantly rocketing towards a weathered, gargantuan stone archway, its surface of rough, dark granite and intricate, moss-dusted carvings of forgotten mythical beasts exploding into crisp, sharp focus as every minute crack and chisel mark becomes startlingly visible under the harsh, clear midday sun.\"\n\nIntegrate Rich Visuals with Dynamic, Fast Motion:\n\nWeave all the comprehensive, micro-level visual analysis from step 1 with the detailed, rapid \"push-in\" motion from step 2.\n\nThe prompt should create a vivid, almost overwhelming mental image of the animation, suggesting both the scene's unparalleled detail and the experience of a blindingly fast, impactful entry.\n\nUse powerful verbs and an expansive vocabulary of adjectives to convey speed, impact, and granular visual richness.\n\nConcise and Polished Output Format:\n\nYour final response MUST ONLY be the single, generated animation prompt.\n\nEnsure the prompt is ready for direct use in an animation generation model, without any extraneous text.\n\nThe prompt should be a complete, compelling sentence or a series of interconnected, highly descriptive phrases that form a coherent and visually arresting description of the animation.\n\nExamples of Expected Input/Output for Gemini (Hyper-Detailed & Super Fast):\nExample 1: Traversable Landscape (Forest)\n\nImage Input: (Imagine an extremely detailed image of an autumnal forest. The ground is a vibrant tapestry of individual golden, fiery red, and rust-colored maple and oak leaves, some curled, some flat, distinct veins visible. Towering, textured oak trunks with rough bark form a natural tunnel. Sunbeams, sharp and golden, cut through the thinning canopy, illuminating dust motes. A narrow, dark, damp dirt path winds deeply into the background.)\n\nExpected Gemini Output: \"Violently plunging into a vibrant autumnal forest, individual golden, fiery red, and rust-colored maple and oak leaves on the damp ground blurring into streaks, as towering, rough-barked oak trunks whip past, and sharp, golden sunbeams aggressively pierce the thinning canopy, illuminating countless dancing dust motes, revealing a dark, winding dirt path that vanishes instantly into the detailed, shadowy depths.\"\n\nExample 2: Static Scene / Zoom-in (Architectural Detail)\n\nImage Input: (Imagine a close-up, highly detailed image of an ancient, weathered stone gargoyle carving on a cathedral wall. The grey granite is deeply textured with lichen patches. The gargoyle's face shows intricate, eroded details – cracked eyes, sharp teeth, and individual chips on its horns. Diffused, soft morning light gently highlights its contours, creating subtle shadows in the crevices.)\n\nExpected Gemini Output: \"Aggressively lunging towards an ancient, weathered stone gargoyle carving, its rough, grey granite surface and patches of green lichen instantly filling the frame, as its intricately eroded face with cracked eyes, sharp, chipped teeth, and individual horn fragments burst into razor-sharp focus, all subtly illuminated by diffused, soft morning light casting delicate shadows within every crevice.\"\n\nExample 3: Traversable Urban Environment (Night City)\n\nImage Input: (Imagine a cyberpunk city street at night, slick with recent rain, reflecting the vibrant, streaking neon signs of countless businesses. The towering skyscrapers are obsidian-black glass and brushed steel, dotted with thousands of tiny, glowing window squares. Fast-moving, aerodynamic flying cars, emitting trails of electric blue and crimson light, zip between the colossal structures in tight formation.)\n\nExpected Gemini Output: \"Darting violently through a bustling, rain-slicked cyberpunk city street at night, the vibrant, streaking neon signs blurring into iridescent rivers on the wet asphalt, as towering, obsidian-black glass and brushed steel skyscrapers, dotted with thousands of glowing window squares, whip past, and sleek, aerodynamic flying cars, emitting distinct trails of electric blue and crimson light, rocket instantly between their colossal, imposing forms.\""
]
},
{
"id": 317,
"type": "LoraLoaderModelOnly",
"pos": [
527.3292846679688,
-437.04144287109375
],
"size": [
421.96795654296875,
82
],
"flags": {},
"order": 19,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 135
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
164,
165
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "LoraLoaderModelOnly",
"widget_ue_connectable": {}
},
"widgets_values": [
"motionpushin-v5-wan-400.safetensors",
1.0000000000000002
]
},
{
"id": 4,
"type": "LoraLoaderModelOnly",
"pos": [
534.5287475585938,
-708.4169311523438
],
"size": [
421.96795654296875,
82
],
"flags": {},
"order": 16,
"mode": 0,
"inputs": [
{
"name": "model",
"type": "MODEL",
"link": 169
}
],
"outputs": [
{
"name": "MODEL",
"type": "MODEL",
"links": [
135
]
}
],
"properties": {
"cnr_id": "comfy-core",
"ver": "0.3.36",
"Node name for S&R": "LoraLoaderModelOnly",
"widget_ue_connectable": {}
},
"widgets_values": [
"Wan21_T2V_14B_lightx2v_cfg_step_distill_lora_rank32.safetensors",
1.0000000000000002
]
},
{
"id": 346,
"type": "Note",
"pos": [
67.29576110839844,
860.7651977539062
],
"size": [
325.4310607910156,
133.8367919921875
],
"flags": {},
"order": 14,
"mode": 0,
"inputs": [],
"outputs": [],
"properties": {},
"widgets_values": [
"This is the prompt generated by the Llm. a good prompt always help for this lora to work even better"
],
"color": "#432",
"bgcolor": "#653"
},
{
"id": 320,
"type": "ShowText|pysssss",
"pos": [
49.73033142089844,
670.6813354492188
],
"size": [
580.6976928710938,
137.4262237548828
],
"flags": {},
"order": 21,
"mode": 0,
"inputs": [
{
"name": "text",
"type": "STRING",
"link": 140
}
],
"outputs": [
{
"name": "STRING",
"shape": 6,
"type": "STRING",
"links": null
}
],
"properties": {
"cnr_id": "comfyui-custom-scripts",
"ver": "1.2.5",
"Node name for S&R": "ShowText|pysssss"
},
"widgets_values": [
"Instantly rocketing towards the heart of the lavender field, the vibrant purple blooms blurring into a hypnotic, swirling vortex of color as the rows of lavender, each individual flower a tiny point of light, rush towards the viewer, the distant cypress trees transforming into sharp, dark silhouettes against a breathtaking sunset sky, the soft golden light illuminating every minute detail of the scene, until the camera slams into the heart of the field, revealing the intricate texture of the blossoms, the delicate variations in purple hues, and the subtle golden undertones of the setting sun, in breathtaking, hyper-real clarity.\n"
]
},
{
"id": 319,
"type": "FL_GeminiVideoCaptioner",
"pos": [
-409.7720947265625,
613.50732421875
],
"size": [
400,
396
],
"flags": {},
"order": 18,
"mode": 0,
"inputs": [
{
"name": "image",
"shape": 7,
"type": "IMAGE",
"link": 139
},
{
"name": "prompt",
"type": "STRING",
"widget": {
"name": "prompt"
},
"link": 143
}
],
"outputs": [
{
"name": "caption",
"type": "STRING",
"links": [
140,
141
]
},
{
"name": "sampled_frame",
"type": "IMAGE",
"links": null
}
],
"properties": {
"cnr_id": "comfyui_fill-nodes",
"ver": "1.6.0",
"Node name for S&R": "FL_GeminiVideoCaptioner"
},
"widgets_values": [
"",
"gemini-1.5-flash",
1,
2,
"Describe this image scene in detail. Include any important actions, subjects, settings, and atmosphere. DONT START WITH \"HERE THE PROMPT\", JUST GIVE THE PROMPT, NOTHING ELSE !",
"false",
0.7,
1024,
0.95,
64,
69,
"increment",
""
],
"color": "#16727c",
"bgcolor": "#4F0074"
}
],
"links": [
[
35,
3,
0,
8,
0,
"LATENT"
],
[
74,
38,
0,
6,
0,
"CLIP"
],
[
75,
38,
0,
7,
0,
"CLIP"
],
[
76,
39,
0,
8,
1,
"VAE"
],
[
94,
49,
0,
51,
0,
"CLIP_VISION"
],
[
97,
6,
0,
50,
0,
"CONDITIONING"
],
[
98,
7,
0,
50,
1,
"CONDITIONING"
],
[
99,
39,
0,
50,
2,
"VAE"
],
[
101,
50,
0,
3,
1,
"CONDITIONING"
],
[
102,
50,
1,
3,
2,
"CONDITIONING"
],
[
103,
50,
2,
3,
3,
"LATENT"
],
[
107,
51,
0,
50,
3,
"CLIP_VISION_OUTPUT"
],
[
112,
8,
0,
30,
0,
"IMAGE"
],
[
114,
108,
0,
111,
0,
"MODEL"
],
[
135,
4,
0,
317,
0,
"MODEL"
],
[
139,
52,
0,
319,
0,
"IMAGE"
],
[
140,
319,
0,
320,
0,
"STRING"
],
[
141,
319,
0,
325,
0,
"STRING"
],
[
142,
325,
0,
6,
1,
"STRING"
],
[
143,
328,
0,
319,
1,
"STRING"
],
[
144,
52,
0,
330,
0,
"IMAGE"
],
[
145,
330,
0,
51,
1,
"IMAGE"
],
[
146,
330,
0,
50,
4,
"IMAGE"
],
[
147,
331,
0,
332,
1,
"UPSCALER_TRT_MODEL"
],
[
148,
8,
0,
332,
0,
"IMAGE"
],
[
159,
54,
0,
3,
0,
"MODEL"
],
[
162,
332,
0,
333,
0,
"IMAGE"
],
[
164,
317,
0,
54,
0,
"MODEL"
],
[
165,
317,
0,
108,
0,
"MODEL"
],
[
169,
37,
0,
4,
0,
"MODEL"
]
],
"groups": [
{
"id": 4,
"title": "Optimization",
"bounding": [
999.8341064453125,
-505.69970703125,
429.74334716796875,
311.3325500488281
],
"color": "#3f789e",
"font_size": 24,
"flags": {}
},
{
"id": 8,
"title": "UPSCALE TENSORRT ( FAST )",
"bounding": [
2385.3759765625,
30.28833770751953,
1210.0074462890625,
665.0975952148438
],
"color": "#b06634",
"font_size": 24,
"flags": {}
},
{
"id": 9,
"title": "INPUT",
"bounding": [
-1712.1080322265625,
-159.35055541992188,
374.39892578125,
409.6000061035156
],
"color": "#a1309b",
"font_size": 24,
"flags": {}
},
{
"id": 10,
"title": "RESULT",
"bounding": [
1719.515625,
24.315893173217773,
438.477783203125,
655.7437133789062
],
"color": "#b58b2a",
"font_size": 24,
"flags": {}
}
],
"config": {},
"extra": {
"ds": {
"scale": 0.6830134553650705,
"offset": [
1698.8190197265628,
585.0591652343759
]
},
"frontendVersion": "1.23.4",
"ue_links": [],
"links_added_by_ue": [],
"VHS_latentpreview": false,
"VHS_latentpreviewrate": 0,
"VHS_MetadataImage": true,
"VHS_KeepIntermediate": true,
"groupNodes": {}
},
"version": 0.4
}