copy and paste this google map to your website or blog!
Press copy button and paste into your blog or website.
(Please switch to 'HTML' mode when posting into your blog. Examples: WordPress Example, Blogger Example)
Wan: Open and Advanced Large-Scale Video Generative Models 💡If the image parameter is configured, it is an Image-to-Video generation; otherwise, it defaults to a Text-to-Video generation 💡Similar to Image-to-Video, the size parameter represents the area of the generated video, with the aspect ratio following that of the original input image Multi-GPU inference using FSDP + DeepSpeed Ulysses
Wan: Open and Advanced Large-Scale Video Generative Models 👍 Multiple Tasks: Wan2 1 excels in Text-to-Video, Image-to-Video, Video Editing, Text-to-Image, and Video-to-Audio, advancing the field of video generation 👍 Visual Text Generation: Wan2 1 is the first video model capable of generating both Chinese and English text, featuring robust text generation that enhances its practical applications
GitHub - lllyasviel FramePack: Lets make video diffusion practical! FramePack is a next-frame (next-frame-section) prediction neural network structure that generates videos progressively FramePack compresses input contexts to a constant length so that the generation workload is invariant to video length FramePack can process a very large number of frames with 13B
hao-ai-lab FastVideo - GitHub FastVideo is a unified framework for accelerated video generation It features a clean, consistent API that works across popular video models, making it easier for developers to author new models and incorporate system- or kernel-level optimizations
DepthAnything Video-Depth-Anything - GitHub This work presents Video Depth Anything based on Depth Anything V2, which can be applied to arbitrarily long videos without compromising quality, consistency, or generalization ability Compared with other diffusion-based models, it enjoys faster inference speed, fewer parameters, and higher
Lightricks LTX-Video: Official repository for LTX-Video - GitHub LTX-Video is the first DiT-based video generation model that can generate high-quality videos in real-time It can generate 30 FPS videos at 1216×704 resolution, faster than it takes to watch them It can generate 30 FPS videos at 1216×704 resolution, faster than it takes to watch them
【EMNLP 2024 】Video-LLaVA: Learning United Visual . . . - GitHub [2024 09 25] 🔥🔥🔥 Our Video-LLaVA has been accepted at EMNLP 2024! We earn the meta score of 4 [2024 07 27] 🔥🔥🔥 A fine-tuned Video-LLaVA focuses on theme exploration, narrative analysis, and character dynamics
GitHub - kijai ComfyUI-WanVideoWrapper Short answer: Unless it's a model feature not available yet on native, you shouldn't Long answer: Due to the complexity of ComfyUI core code, and my lack of coding experience, in many cases it's far easier and faster to implement new models and features to a standalone wrapper, so this is a way to test things relatively quickly