copy and paste this google map to your website or blog!
Press copy button and paste into your blog or website.
(Please switch to 'HTML' mode when posting into your blog. Examples: WordPress Example, Blogger Example)
Wan: Open and Advanced Large-Scale Video Generative Models 👍 Multiple Tasks: Wan2 1 excels in Text-to-Video, Image-to-Video, Video Editing, Text-to-Image, and Video-to-Audio, advancing the field of video generation 👍 Visual Text Generation: Wan2 1 is the first video model capable of generating both Chinese and English text, featuring robust text generation that enhances its practical applications
GitHub - lllyasviel FramePack: Lets make video diffusion practical! FramePack is a next-frame (next-frame-section) prediction neural network structure that generates videos progressively FramePack compresses input contexts to a constant length so that the generation workload is invariant to video length FramePack can process a very large number of frames with 13B
Video-R1: Reinforcing Video Reasoning in MLLMs - GitHub Video-R1 significantly outperforms previous models across most benchmarks Notably, on VSI-Bench, which focuses on spatial reasoning in videos, Video-R1-7B achieves a new state-of-the-art accuracy of 35 8%, surpassing GPT-4o, a proprietary model, while using only 32 frames and 7B parameters
Lightricks LTX-Video: Official repository for LTX-Video - GitHub LTX-Video is the first DiT-based video generation model that can generate high-quality videos in real-time It can generate 30 FPS videos at 1216×704 resolution, faster than it takes to watch them It can generate 30 FPS videos at 1216×704 resolution, faster than it takes to watch them
GitHub - kijai ComfyUI-WanVideoWrapper Short answer: Unless it's a model feature not available yet on native, you shouldn't Long answer: Due to the complexity of ComfyUI core code, and my lack of coding experience, in many cases it's far easier and faster to implement new models and features to a standalone wrapper, so this is a way to test things relatively quickly
GitHub - visomaster VisoMaster: Powerful Easy-to-Use Video Face . . . Live Playback: See processed video in real-time before saving; Face Embeddings: Use multiple source faces for better accuracy similarity; Live Swapping via Webcam: Stream to virtual camera for Twitch, YouTube, Zoom, etc User-Friendly Interface: Intuitive and easy to use; Video Markers: Adjust settings per frame for precise results
hkchengrex MMAudio - GitHub Supports video-to-audio and text-to-audio synthesis You can also try experimental image-to-audio synthesis which duplicates the input image to a video for processing This might be interesting to some but it is not something MMAudio has been trained for Use port forwarding (e g , ssh -L 7860:localhost:7860 server) if necessary
HunyuanVideo: A Systematic Framework For Large Video . . . - GitHub We present HunyuanVideo, a novel open-source video foundation model that exhibits performance in video generation that is comparable to, if not superior to, leading closed-source models In order to train HunyuanVideo model, we adopt several key technologies for model learning, including data