This is awesome.
Tune-A-Video is a new method that allows one-shot tuning of image diffusion models for Text-to-Video generation.
π Paper: arxiv.org
βοΈ Project: tuneavideo.github.io
π οΈ Code: Coming Soon
πMore examples and TLDR below
Tune-A-Video is a new method that allows one-shot tuning of image diffusion models for Text-to-Video generation.
π Paper: arxiv.org
βοΈ Project: tuneavideo.github.io
π οΈ Code: Coming Soon
πMore examples and TLDR below
TLDR: Researchers extended text-to-image diffusion to generate multiple images. To learn continuous motion, they propose Tune-A-Video with a tailored Sparse-Causal Attention, which generates videos from text prompts via an efficient one-shot tuning of pretrained T2I models.
Authors: Jay Zhangjie Wu,@ge_yixiao, Xintao Wang, Stan Weixian Lei, @YuchaoGu, Wynne Hsu, Ying Shan, Xiaohu Qie, @MikeShou1
Loading suggestions...