shridharathi/van-gogh-vid 🔢🖼️❓📝 → 🖼️
About
Make your videos van gogh-esque
Example Output
Prompt:
"GOGH style, painting of a man walking in a field with trees and sunlight"
Output
Performance Metrics
203.91s
Prediction Time
252.24s
Total Time
All Input Parameters
{
"frames": 81,
"prompt": "GOGH style, painting of a man walking in a field with trees and sunlight",
"fast_mode": "Balanced",
"resolution": "480p",
"aspect_ratio": "16:9",
"sample_shift": 8,
"sample_steps": 30,
"negative_prompt": "",
"lora_strength_clip": 1,
"sample_guide_scale": 5,
"lora_strength_model": 1
}
Input Parameters
- seed
- Set a seed for reproducibility. Random by default.
- image
- Image to use as a starting frame for image to video generation.
- frames
- The number of frames to generate (1 to 5 seconds)
- prompt (required)
- Text prompt for video generation
- fast_mode
- Speed up generation with different levels of acceleration. Faster modes may degrade quality somewhat. The speedup is dependent on the content, so different videos may see different speedups.
- resolution
- The resolution of the video. 720p is not supported for 1.3b.
- aspect_ratio
- The aspect ratio of the video. 16:9, 9:16, 1:1, etc.
- sample_shift
- Sample shift factor
- sample_steps
- Number of generation steps. Fewer steps means faster generation, at the expensive of output quality. 30 steps is sufficient for most prompts
- negative_prompt
- Things you do not want to see in your video
- replicate_weights
- Replicate LoRA weights to use. Leave blank to use the default weights.
- lora_strength_clip
- Strength of the LORA applied to the CLIP model. 0.0 is no LORA.
- sample_guide_scale
- Higher guide scale makes prompt adherence better, but can reduce variation
- lora_strength_model
- Strength of the LORA applied to the model. 0.0 is no LORA.
Output Schema
Output
Example Execution Logs
Random seed set to: 578857622
2025-03-28T01:03:35Z | INFO | [ Initiating ] chunk_size=150M dest=/tmp/tmpxptoopvc/weights url=https://replicate.delivery/xezq/KgMA5f1XAqVuW672nNaxVDLb7KfaETMfFCnMHzCJ1R3hqy5oA/trained_model.tar
2025-03-28T01:03:38Z | INFO | [ Complete ] dest=/tmp/tmpxptoopvc/weights size="307 MB" total_elapsed=3.163s url=https://replicate.delivery/xezq/KgMA5f1XAqVuW672nNaxVDLb7KfaETMfFCnMHzCJ1R3hqy5oA/trained_model.tar
Checking inputs
====================================
Checking weights
✅ umt5_xxl_fp16.safetensors exists in ComfyUI/models/text_encoders
✅ 14b_64a1c2e3ddb7864e8e05b8d6455d2865.safetensors exists in loras directory
✅ wan_2.1_vae.safetensors exists in ComfyUI/models/vae
⏳ Downloading wan2.1_t2v_14B_bf16.safetensors to ComfyUI/models/diffusion_models
✅ wan2.1_t2v_14B_bf16.safetensors downloaded to ComfyUI/models/diffusion_models in 14.40s, size: 27253.24MB
====================================
Running workflow
[ComfyUI] got prompt
Executing node 39, title: Load VAE, class type: VAELoader
[ComfyUI] Using pytorch attention in VAE
[ComfyUI] Using pytorch attention in VAE
[ComfyUI] VAE load device: cuda:0, offload device: cpu, dtype: torch.bfloat16
Executing node 40, title: EmptyHunyuanLatentVideo, class type: EmptyHunyuanLatentVideo
Executing node 38, title: Load CLIP, class type: CLIPLoader
[ComfyUI] CLIP/text encoder model load device: cuda:0, offload device: cpu, current: cpu, dtype: torch.float16
[ComfyUI] Requested to load WanTEModel
Executing node 7, title: CLIP Text Encode (Negative Prompt), class type: CLIPTextEncode
[ComfyUI] loaded completely 141327.4875 10835.4765625 True
Executing node 37, title: Load Diffusion Model, class type: UNETLoader
[ComfyUI] model weight dtype torch.float16, manual cast: None
[ComfyUI] model_type FLOW
Executing node 54, title: WanVideo Tea Cache (native), class type: WanVideoTeaCacheKJ
Executing node 49, title: Load LoRA, class type: LoraLoader
[ComfyUI] Requested to load WanTEModel
Executing node 6, title: CLIP Text Encode (Positive Prompt), class type: CLIPTextEncode
[ComfyUI] loaded completely 141069.4875 10835.4765625 True
Executing node 48, title: ModelSamplingSD3, class type: ModelSamplingSD3
Executing node 3, title: KSampler, class type: KSampler
[ComfyUI] Requested to load WAN21
[ComfyUI] loaded completely 124343.96281542968 27251.406372070312 True
[ComfyUI]
[ComfyUI] 0%| | 0/30 [00:00<?, ?it/s]
[ComfyUI] 3%|▎ | 1/30 [00:06<03:21, 6.94s/it]
[ComfyUI] 7%|▋ | 2/30 [00:16<03:53, 8.35s/it]
[ComfyUI] 10%|█ | 3/30 [00:25<03:57, 8.81s/it]
[ComfyUI] TeaCache: Initialized
[ComfyUI]
[ComfyUI] 13%|█▎ | 4/30 [00:38<04:26, 10.25s/it]
[ComfyUI] 20%|██ | 6/30 [00:47<02:57, 7.41s/it]
[ComfyUI] 27%|██▋ | 8/30 [00:57<02:18, 6.32s/it]
[ComfyUI] 33%|███▎ | 10/30 [01:07<01:55, 5.77s/it]
[ComfyUI] 40%|████ | 12/30 [01:17<01:38, 5.45s/it]
[ComfyUI] 47%|████▋ | 14/30 [01:26<01:24, 5.26s/it]
[ComfyUI] 53%|█████▎ | 16/30 [01:36<01:11, 5.13s/it]
[ComfyUI] 60%|██████ | 18/30 [01:46<01:00, 5.05s/it]
[ComfyUI] 67%|██████▋ | 20/30 [01:56<00:49, 4.99s/it]
[ComfyUI] 73%|███████▎ | 22/30 [02:05<00:39, 4.95s/it]
[ComfyUI] 80%|████████ | 24/30 [02:15<00:29, 4.98s/it]
[ComfyUI] 83%|████████▎ | 25/30 [02:15<00:20, 4.11s/it]
[ComfyUI] 87%|████████▋ | 26/30 [02:25<00:21, 5.26s/it]
[ComfyUI] 90%|█████████ | 27/30 [02:25<00:12, 4.10s/it]
[ComfyUI] 93%|█████████▎| 28/30 [02:35<00:10, 5.46s/it]
[ComfyUI] 100%|██████████| 30/30 [02:45<00:00, 5.22s/it]
[ComfyUI] 100%|██████████| 30/30 [02:45<00:00, 5.51s/it]
[ComfyUI] Requested to load WanVAE
Executing node 8, title: VAE Decode, class type: VAEDecode
[ComfyUI] loaded completely 98169.38668441772 242.02829551696777 True
Executing node 50, title: Video Combine 🎥🅥🅗🅢, class type: VHS_VideoCombine
[ComfyUI] Prompt executed in 186.12 seconds
outputs: {'50': {'gifs': [{'filename': 'R8_Wan_00001.mp4', 'subfolder': '', 'type': 'output', 'format': 'video/h264-mp4', 'frame_rate': 16.0, 'workflow': 'R8_Wan_00001.png', 'fullpath': '/tmp/outputs/R8_Wan_00001.mp4'}]}}
====================================
R8_Wan_00001.png
R8_Wan_00001.mp4
Version Details
- Version ID
bee70f3c8f0db29784b12b5129186eaa9607104308f592c5f07274cac7acbd2a- Version Created
- March 27, 2025