The aim of fusing infrared and visible images is to achieve high-quality images by enhancing textural details and obtaining complementary benefits. Since the details of the visible images are not obvious in low light, it is difficult for the current fusion methods to complete the complementary contours and texture details. With the intention of addressing the challenge of poor quality of infrared and visible light fusion images under low light conditions, a novel fusion method for infrared and visible light is presented in this study utilizing generative adversarial networks (referred to as UFIVL). Specifically, based on the existing densely connected decoder, pruning is introduced to reduce the network complexity without quality loss. A new overall optimization objective includes the adaptive limit contrast histogram equalization loss and the joint gradient loss are designed to deal with the defects of high contrast and brightness loss of the fused image, and the difficulty of capturing detailed features in low light scenes, respectively. Experimental results on LLVIP datasets show that compared with other state-of-the-art methods, the fused image generated by the proposed method has better subjective and objective performances.
Modal translation between multimodal images is an effective complementary scheme when images with some certain modal are difficult to obtain. Since pixel level image modal translation method can obtain the images with the highest quality, it has become a research hotspot in recent years. Generative adversarial network (GAN) is a network for image generation, due to the complex structure of GAN and the complexity of the image generation task, the training results of GAN are not stable. In this paper, on the basis of U-nets, the dense block is used to increase the feature information in the subsampling encoding and up-sampling decoding operation, so as to reduce the loss of information and obtain higher quality images. At the same time, the dense long connection is introduced to connect the encoding and decoding operations of the same stage, so that the network can effectively combine the features at low and high level, and improve the performance of the network. Experimental results show that the proposed method is effective in modal translation of multimodal images, and the image quality is better than some state-of-the-art methods.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.