Javascript must be enabled to continue!
ITrans: Generative Image Inpainting with Transformers (ChinaMM)
View through CrossRef
Abstract
Despite significant improvements, convolutional neural network (CNN) based methods are struggling with handling long-range global image dependencies due to their limited receptive fields, leading to an unsatisfactory inpainting performance under complicated scenarios. To address this issue, we propose the Inpainting Transformer (ITrans) network, which combines the power of both self-attention and convolution operations. The ITrans network augments convolutional encoder-decoder structure with two novel designs, \ie, the Global and Local Transformers. The Global Transformer aggregates high-level image context from the encoder in a global perspective, and propagates the encoded global representation to the decoder in a multi-scale manner. Meanwhile, the Local Transformer is intended to extract low-level image details inside the local neighborhood at a reduced computational overhead. By incorporating the above two Transformers, ITrans is capable of both global relationship modeling and local details encoding, which is essential for hallucinating perceptually realistic images. Extensive experiments demonstrate that the proposed ITrans network outperforms favorably against state-of-the-art inpainting methods both quantitatively and qualitatively.
Research Square Platform LLC
Title: ITrans: Generative Image Inpainting with Transformers (ChinaMM)
Description:
Abstract
Despite significant improvements, convolutional neural network (CNN) based methods are struggling with handling long-range global image dependencies due to their limited receptive fields, leading to an unsatisfactory inpainting performance under complicated scenarios.
To address this issue, we propose the Inpainting Transformer (ITrans) network, which combines the power of both self-attention and convolution operations.
The ITrans network augments convolutional encoder-decoder structure with two novel designs, \ie, the Global and Local Transformers.
The Global Transformer aggregates high-level image context from the encoder in a global perspective, and propagates the encoded global representation to the decoder in a multi-scale manner.
Meanwhile, the Local Transformer is intended to extract low-level image details inside the local neighborhood at a reduced computational overhead.
By incorporating the above two Transformers, ITrans is capable of both global relationship modeling and local details encoding, which is essential for hallucinating perceptually realistic images.
Extensive experiments demonstrate that the proposed ITrans network outperforms favorably against state-of-the-art inpainting methods both quantitatively and qualitatively.
Related Results
Virtual Inpainting for Dazu Rock Carvings Based on a Sample Dataset
Virtual Inpainting for Dazu Rock Carvings Based on a Sample Dataset
Numerous image inpainting algorithms are guided by a basic assumption that the known region in the original image itself can provide sufficient prior information for the guess reco...
Diversity-Generated Image Inpainting with Style Extraction
Diversity-Generated Image Inpainting with Style Extraction
The latest methods based on deep learning have achieved amazing results regarding the complex work of inpainting large missing areas in an image. This type of method generally atte...
MD-GAN: Multi-Scale Diversity GAN for Large Masks Inpainting
MD-GAN: Multi-Scale Diversity GAN for Large Masks Inpainting
Image inpainting approaches have made considerable progress with the assistance of generative adversarial networks (GANs) recently. However, current inpainting methods are incompet...
Ancient mural inpainting via structure information guided two-branch model
Ancient mural inpainting via structure information guided two-branch model
AbstractAncient murals are important cultural heritages for our exploration of ancient civilizations and are of great research value. Due to long-time exposure to the environment, ...
Iterative Geometry-Aware Cross Guidance Network for Stereo Image Inpainting
Iterative Geometry-Aware Cross Guidance Network for Stereo Image Inpainting
Currently, single image inpainting has achieved promising results based on deep convolutional neural networks. However, inpainting on stereo images with missing regi...
On the Remote Calibration of Instrumentation Transformers: Influence of Temperature
On the Remote Calibration of Instrumentation Transformers: Influence of Temperature
The remote calibration of instrumentation transformers is theoretically possible using synchronous measurements across a transmission line with a known impedance and a local set of...
Double Exposure
Double Exposure
I. Happy Endings
Chaplin’s Modern Times features one of the most subtly strange endings in Hollywood history. It concludes with the Tramp (Chaplin) and the Gamin (Paulette Godda...
Increased Transformer Availability and Reliability
Increased Transformer Availability and Reliability
Abstract
Transformers are important components of the High Voltage electrical grid and electrical power installation in industrial plants such as the petroleum indus...

