Trending repositories for topic video-generation
HunyuanVideo: A Systematic Framework For Large Video Generation Model
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
Text and image to video generation: Kandinsky 4.0 (2024)
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
Identity-Preserving Text-to-Video Generation by Frequency Decomposition
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
Code of Pyramidal Flow Matching for Efficient Video Generative Modeling
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
A repository for organizing papers, codes and other resources related to Virtual Try-on Models
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
[ICCV 2023 Oral] Text-to-Image Diffusion Models are Zero-Shot Video Generators
Text and image to video generation: Kandinsky 4.0 (2024)
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
A repository for organizing papers, codes and other resources related to Virtual Try-on Models
HunyuanVideo: A Systematic Framework For Large Video Generation Model
Pytorch implementation of "Genie: Generative Interactive Environments", Bruce et al. (2024).
Identity-Preserving Text-to-Video Generation by Frequency Decomposition
📚 Collection of awesome generation acceleration resources.
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
To support and further the research in the field of portrait animation , we are excited to launch PhotoPoster, an open project for pose-driven image generation.
Adaptive Caching for Faster Video Generation with Diffusion Transformers
[NeurIPS D&B Track 2024] Official implementation of HumanVid
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
The Dawn of Video Generation: Preliminary Explorations with SORA-like Models
HunyuanVideo: A Systematic Framework For Large Video Generation Model
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
Text and image to video generation: Kandinsky 4.0 (2024)
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.
Identity-Preserving Text-to-Video Generation by Frequency Decomposition
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
Code of Pyramidal Flow Matching for Efficient Video Generative Modeling
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
The Dawn of Video Generation: Preliminary Explorations with SORA-like Models
Official code of "Imagine360: Immersive 360 Video Generation from Perspective Anchor"
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
A repository for organizing papers, codes and other resources related to Virtual Try-on Models
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
Official code of "Imagine360: Immersive 360 Video Generation from Perspective Anchor"
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
HunyuanVideo: A Systematic Framework For Large Video Generation Model
A repository for organizing papers, codes and other resources related to Virtual Try-on Models
The Dawn of Video Generation: Preliminary Explorations with SORA-like Models
Identity-Preserving Text-to-Video Generation by Frequency Decomposition
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
Pytorch implementation of "Genie: Generative Interactive Environments", Bruce et al. (2024).
📚 Collection of awesome generation acceleration resources.
JAX reimplementation of the DeepMind paper "Genie: Generative Interactive Environments"
Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
Adaptive Caching for Faster Video Generation with Diffusion Transformers
A one-stop library to standardize the inference and evaluation of all the conditional video generation models.
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
Official code of "Imagine360: Immersive 360 Video Generation from Perspective Anchor"
HunyuanVideo: A Systematic Framework For Large Video Generation Model
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
Identity-Preserving Text-to-Video Generation by Frequency Decomposition
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
Code of Pyramidal Flow Matching for Efficient Video Generative Modeling
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.
Text and image to video generation: Kandinsky 4.0 (2024)
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
Official code of "Imagine360: Immersive 360 Video Generation from Perspective Anchor"
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
The Dawn of Video Generation: Preliminary Explorations with SORA-like Models
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
Official code of "Imagine360: Immersive 360 Video Generation from Perspective Anchor"
Official implementation of the paper “MagicDriveDiT: High-Resolution Long Video Generation for Autonomous Driving with Adaptive Control”
HunyuanVideo: A Systematic Framework For Large Video Generation Model
A repository for organizing papers, codes and other resources related to Virtual Try-on Models
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
Official implementation of the paper "Koala-36M: A Large-scale Video Dataset Improving Consistency between Fine-grained Conditions and Video Content".
📚 Collection of awesome generation acceleration resources.
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
The Dawn of Video Generation: Preliminary Explorations with SORA-like Models
Pytorch implementation of "Genie: Generative Interactive Environments", Bruce et al. (2024).
Adaptive Caching for Faster Video Generation with Diffusion Transformers
JAX reimplementation of the DeepMind paper "Genie: Generative Interactive Environments"
Official repository for "iVideoGPT: Interactive VideoGPTs are Scalable World Models" (NeurIPS 2024), https://arxiv.org/abs/2405.15223
[NeurIPS 2024] Animate3D: Animating Any 3D Model with Multi-view Video Diffusion
Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance
Code of Pyramidal Flow Matching for Efficient Video Generative Modeling
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
Official implementations for paper: DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
MiniSora: A community aims to explore the implementation path and future development direction of Sora.
Code for Paper "UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation".
[AAAI 2025] Follow-Your-Click: This repo is the official implementation of "Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts"
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
[SIGGRAPH ASIA 2024 TCS] AnimateLCM: Computation-Efficient Personalized Style Video Generation without Personalized Video Data
Identity-Preserving Text-to-Video Generation by Frequency Decomposition
Official codes of VEnhancer: Generative Space-Time Enhancement for Video Generation
A Collection of Papers and Codes for CVPR2024/ECCV2024 AIGC
text and image to video generation: CogVideoX (2024) and CogVideo (ICLR 2023)
Champ: Controllable and Consistent Human Image Animation with 3D Parametric Guidance
HunyuanVideo: A Systematic Framework For Large Video Generation Model
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors
A curated list of recent diffusion models for video generation, editing, restoration, understanding, etc.
Code of Pyramidal Flow Matching for Efficient Video Generative Modeling
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
Official implementations for paper: DreamTalk: When Expressive Talking Head Generation Meets Diffusion Probabilistic Models
VideoCrafter2: Overcoming Data Limitations for High-Quality Video Diffusion Models
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
MiniSora: A community aims to explore the implementation path and future development direction of Sora.
Code for Paper "UniAnimate: Taming Unified Video Diffusion Models for Consistent Human Image Animation".
[AAAI 2025] Follow-Your-Click: This repo is the official implementation of "Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts"
Fine-Grained Open Domain Image Animation with Motion Guidance
Quantized Attention that achieves speedups of 2.1-3.1x and 2.7-5.1x compared to FlashAttention2 and xformers, respectively, without lossing end-to-end metrics across various models.
[ICLR24] Official implementation of the paper “MagicDrive: Street View Generation with Diverse 3D Geometry Control”
[SIGGRAPH ASIA 2024 TCS] AnimateLCM: Computation-Efficient Personalized Style Video Generation without Personalized Video Data
A python tool that uses GPT-4, FFmpeg, and OpenCV to automatically analyze videos, extract the most interesting sections, and crop them for an improved viewing experience.
MagicTime: Time-lapse Video Generation Models as Metamorphic Simulators
MuseV: Infinite-length and High Fidelity Virtual Human Video Generation with Visual Conditioned Parallel Denoising
[AAAI 2025] Follow-Your-Click: This repo is the official implementation of "Follow-Your-Click: Open-domain Regional Image Animation via Short Prompts"
MiniSora: A community aims to explore the implementation path and future development direction of Sora.
PhysGen: Rigid-Body Physics-Grounded Image-to-Video Generation (ECCV 2024)
Sora AI Awesome List – Your go-to resource hub for all things Sora AI, OpenAI's groundbreaking model for crafting realistic scenes from text. Explore a curated collection of articles, videos, podcasts...
[ARXIV'24] SynCamMaster: Synchronizing Multi-Camera Video Generation from Diverse Viewpoints
ConsistI2V: Enhancing Visual Consistency for Image-to-Video Generation (TMLR 2024)
To support and further the research in the field of portrait animation , we are excited to launch PhotoPoster, an open project for pose-driven image generation.
High-Quality Human Motion Video Generation with Confidence-aware Pose Guidance
Generative Camera Dolly: Extreme Monocular Dynamic Novel View Synthesis (ECCV 2024 Oral) - Official Implementation
InteractiveVideo: User-Centric Controllable Video Generation with Synergistic Multimodal Instructions
[ECCV 2024 Oral] EDTalk - Official PyTorch Implementation
[CVPR 2024] On the Content Bias in Fréchet Video Distance
📚 Collection of awesome generation acceleration resources.
Official code of "Imagine360: Immersive 360 Video Generation from Perspective Anchor"
[ECCV 2024, Oral] DynamiCrafter: Animating Open-domain Images with Video Diffusion Priors