论文标题

使用语义指导插入互动图像

Interactive Image Inpainting Using Semantic Guidance

论文作者

Yu, Wangbo, Du, Jinhao, Liu, Ruixin, Li, Yixuan, zhu, Yuesheng

论文摘要

图像介绍方法在深层神经网络的帮助下取得了重大进展。但是,现有方法主要集中于利用神经网络学到的先验分布来产生单个介绍结果或进一步产生多种解决方案,而该解决方案没有得到很好的研究。本文开发了一种新颖的图像介绍方法,使用户可以通过自己的喜好或内存来自定义介入结果。具体而言,我们的方法由两个阶段组成,这些阶段利用了神经网络的先前和用户指导来共同注册损坏的图像。在第一阶段,部署了基于新颖的外部空间注意机制的自动编码器,以产生损坏的图像的重建特征,并具有粗糙的镶嵌结果,可提供语义掩码作为用户交互的媒介。在第二阶段,采用了重建功能的语义解码器,以综合以用户自定义的语义掩码为指导的精细填充结果,以便最终的插入结果将与用户的指导共享相同的内容,而在第一个阶段保留了重建的纹理和颜色。广泛的实验证明了我们方法在质量和可控性方面的优势。

Image inpainting approaches have achieved significant progress with the help of deep neural networks. However, existing approaches mainly focus on leveraging the priori distribution learned by neural networks to produce a single inpainting result or further yielding multiple solutions, where the controllability is not well studied. This paper develops a novel image inpainting approach that enables users to customize the inpainting result by their own preference or memory. Specifically, our approach is composed of two stages that utilize the prior of neural network and user's guidance to jointly inpaint corrupted images. In the first stage, an autoencoder based on a novel external spatial attention mechanism is deployed to produce reconstructed features of the corrupted image and a coarse inpainting result that provides semantic mask as the medium for user interaction. In the second stage, a semantic decoder that takes the reconstructed features as prior is adopted to synthesize a fine inpainting result guided by user's customized semantic mask, so that the final inpainting result will share the same content with user's guidance while the textures and colors reconstructed in the first stage are preserved. Extensive experiments demonstrate the superiority of our approach in terms of inpainting quality and controllability.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源