We introduce SeedEdit, a diffusion model that is able to revise a given image with any text prompt. In our perspective, the key to such a task is to obtain an optimal balance between maintaining the original image, i.e. image reconstruction, and generating a new image, i.e. image re-generation. To this end, we start from a weak generator (text-to-image model) that creates diverse pairs between such two directions and gradually align it into a strong image editor that well balances between the two tasks. SeedEdit can achieve more diverse and stable editing capability over prior image editing methods, enabling sequential revision over images generated by diffusion models.
翻译:我们提出了SeedEdit,一种能够根据任意文本提示修改给定图像的扩散模型。我们认为,此类任务的关键在于保持原始图像(即图像重建)与生成新图像(即图像再生)之间的最佳平衡。为此,我们从生成多样化图像对(重建与再生方向)的弱生成器(文本到图像模型)出发,逐步将其对齐为能良好平衡这两项任务的强图像编辑器。相较于现有图像编辑方法,SeedEdit能够实现更多样且稳定的编辑能力,支持对扩散模型生成图像进行序列化修订。