BDMFuse: Multi-scale network fusion for infrared and visible images based on base and detail features
CSTR:
Author:
Affiliation:

1.College of Information and Management Science, Henan Agricultural University, Zhengzhou 450046, China;2.NOVA Information Management School, Universidade Nova de Lisboa, Lisboa1070-312, Portugal

Clc Number:

TP391.4

Fund Project:

Supported by the Henan Province Key Research and Development Project (231111211300), the Central Government of Henan Province Guides Local Science and Technology Development Funds (Z20231811005), Henan Province Key Research and Development Project (231111110100), Henan Provincial Outstanding Foreign Scientist Studio (GZS2024006), and Henan Provincial Joint Fund for Scientific and Technological Research and Development Plan (Application and Overcoming Technical Barriers) (242103810028)

  • Article
  • |
  • Figures
  • |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • |
  • Materials
  • |
  • Comments
    Abstract:

    The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images. To meet these requirements, an autoencoder-based method for infrared and visible image fusion is proposed. The encoder designed according to the optimization objective consists of a base encoder and a detail encoder, which is used to extract low-frequency and high-frequency information from the image. This extraction may lead to some information not being captured, so a compensation encoder is proposed to supplement the missing information. Multi-scale decomposition is also employed to extract image features more comprehensively. The decoder combines low-frequency, high-frequency and supplementary information to obtain multi-scale features. Subsequently, the attention strategy and fusion module are introduced to perform multi-scale fusion for image reconstruction. Experimental results on three datasets show that the fused images generated by this network effectively retain salient targets while being more consistent with human visual perception.

    Reference
    Related
    Cited by
Get Citation

SI Hai-Ping, ZHAO Wen-Rui, LI Ting-Ting, LI Fei-Tao, FERNADO Bacao, SUN Chang-Xia, LI Yan-Ling. BDMFuse: Multi-scale network fusion for infrared and visible images based on base and detail features[J]. Journal of Infrared and Millimeter Waves,2025,44(2):275~284

Copy
Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:July 25,2024
  • Revised:February 09,2025
  • Adopted:September 13,2024
  • Online: February 08,2025
  • Published: April 25,2025
Article QR Code