For recent diffusion-based generative models, maintaining consistent content across a series of generated images, especially those containing subjects and complex details, presents a significant challenge. In this paper, we propose a new way of self-attention calculation, termed Consistent Self-Attention, that significantly boosts the consistency between the generated images and augments prevalent pretrained diffusion-based text-to-image models in a zero-shot manner. To extend our method to long-range video generation, we further introduce a novel semantic space temporal motion prediction module, named Semantic Motion Predictor. It is trained to estimate the motion conditions between two provided images in the semantic spaces. This module converts the generated sequence of images into videos with smooth transitions and consistent subjects that are significantly more stable than the modules based on latent spaces only, especially in the context of long video generation. By merging these two novel components, our framework, referred to as StoryDiffusion, can describe a text-based story with consistent images or videos encompassing a rich variety of contents. The proposed StoryDiffusion encompasses pioneering explorations in visual story generation with the presentation of images and videos, which we hope could inspire more research from the aspect of architectural modifications. Our code is made publicly available at https://github.com/HVision-NKU/StoryDiffusion.
翻译:对于当前基于扩散的生成模型而言,在包含主体与复杂细节的一系列生成图像中保持内容一致性是一项重大挑战。本文提出了一种名为一致性自注意力机制(Consistent Self-Attention)的自注意力计算新方法,该方法能显著提升生成图像间的一致性,并以零样本方式增强现有预训练扩散文本到图像模型。为将所提方法扩展至长程视频生成,我们进一步引入了一种新型语义空间时序运动预测模块——语义运动预测器(Semantic Motion Predictor)。该模块经过训练,可在语义空间中估计两幅给定图像间的运动条件,从而将生成的图像序列转换为具有平滑过渡且主体一致的视频,其稳定性显著优于仅基于隐空间的模块,尤其在长视频生成场景下表现更为突出。通过融合这两项创新组件,我们的框架(称为StoryDiffusion)能够用一致的图像或视频描述基于文本的故事,涵盖丰富多样的内容。所提出的StoryDiffusion在视觉故事生成领域开展了开创性探索,以图像与视频形式呈现内容,期望能从架构改进层面激发更多研究。我们的代码已开源至https://github.com/HVision-NKU/StoryDiffusion。