Contents
In many computer vision applications (e.g. robot motion and medical imaging) there is a need to integrate relevant information from multiple images into a single image. Such image fusion will provide higher reliability, accuracy, and data quality.
Multiview fusion improves the image with higher resolution and also recovers the 3D representation of a scene. Multimodal fusion combines images from different sensors and is referred to as multi-sensor fusion. Its main applications include medical imagery, surveillance, and security.
About us: ProX PC provides a robust end-to-end computer vision solution – Our software helps several leading organizations start with computer vision and implement deep learning models efficiently with minimal overhead for various downstream tasks. Get a demo here.
Levels of Image Fusion
Engineers perform Image Fusion (IF) at three levels based on the stage of fusion accomplishment.
Levels of Image Fusion
Types of Image Fusion
Single-sensor IF
Single-sensor image fusion captures the real world as a sequence of images. The algorithm combines a set of images and generates a new image with optimal information content. E.g. in different lighting conditions, a human operator may not be able to detect objects but highlights them in the resultant fused image.
The drawbacks of this method are the limitations of the imaging sensor that is used in some sensing areas. The conditions in which the sensor capability restricts the system functions (dynamic range, resolution, etc.). For example, some sensors are good for illuminated environments (daylight) but are not suitable for night and fog conditions.
Multi-sensor IF
A multi-sensor image fusion method merges the images from several sensors to form a composite image. E.g. an infrared camera and a separate digital camera produce their individual images and by merging, the final fused image is produced. This approach overcomes the single-sensor problems.
Multi-sensory Image Fusion
This method generates the merged information from several images. The digital camera is suitable for daylight conditions; the infrared camera is good in weakly illuminated environments. So the method has applications in the military and also in object detection, robotics, and medical imaging.
Multiview IF
In this method, images have multiple or different views at the same time. This method utilizes images from different conditions like visible, infrared, multispectral, and remote sensing. Common methods of image fusion include object-level fusion, weighted pixel fusion, and fusion in the transform domain.
Multi-focus IF
This method processes images from 3D views with their focal length. It splits the original image into regions so that every region is in focus for at least one channel of the image.
How to Implement Image Fusion?
Researchers implement image fusion in multiple ways and here we present the most common methods.
Convolutional Neural Network
Zhang et al. (2021) created a CNN-based fusion framework to extract features and reconstruct images by using a carefully designed loss function. They utilized CNN as part of the overall fusion framework to perform activity-level monitoring and feature integration.
In their case of CNN for fusion, they combined loss function with classified CNN to perform medical IF. In addition, they embedded the fusion layer in the training process. Therefore, CNN reduces the constraints caused by manually designed fusion rules (maximum, minimum, or average).
IF Implementation by CNN
Also, the researchers introduced other approaches:
Multiscale Transformation
Ma et al. (2023) performed the fusion process by using multiscale transformation:
IF Implementation by Multiscale Transformation
Sparse Representation Model for IF
Compared to traditional multiscale transform, sparse representation has two main differences. The multiscale fusion method uses a preset basis function, which ignores some important features of the source image. The sparse representation learns over a complete feature set, which can better express and extract images.
In addition, the multiscale transform-based fusion method decomposes images into multiple layers, but the requirements for noise and registration are quite strict. The sparse representation uses a sliding window technique to segment the image into multiple overlapping segments, which improves robustness.
Sparse Representation Model for IF
Because of the sliding window, there’s an overlapping small block, which lowers the operational efficiency of the algorithm.
Applications of Image Fusion
The four main IF use cases are:
Robotic Vision
The robotic motion utilizes the fusion of infrared and visible images. Robots use infrared images to distinguish the target from the background, because of the difference in thermal radiation. Therefore, the illumination and weather conditions do not affect the fusion. However, infrared images don’t provide texture detail.
For their computer vision tasks, robots utilize visible light images. Because of the influence of the data collection environment, visible images may not show important targets. Infrared and visible light fusion methods overcome this drawback of a single image, thus extracting information.
Robotic vision – Amazon humanoid robot
The fusion images are usually clearer than the infrared images. In addition, robots perform a fusion of visible and infrared images, such as for autonomous driving and face recognition.
Medical Imagery
Today, medical imagery generates various types of medical images to help doctors diagnose diseases or injuries. Each type of image has its specific intensity. Therefore, IF has a high clinical application in the field of medical imaging modalities.
Medical imagery researchers combine redundant information and related information from different medical images, to create fused medical images. Thus they provide quality information-inspired image diagnosis for their medical examinations.
Image Fusion in Medical Imagery
The figure shows an example of image fusion for medical diagnostics by combining Computed Tomography (CT) and MRI. The data comes from a brain image dataset of combined tomography and magnetic resonance imaging (MedPix dataset).
Doctors use CT to analyze bone structures with high-spatial domain resolution, and MRI to detect soft tissues, such as the heart, eyes, and brain. MRI and CT are combined with image fusion technology to increase accuracy and medical applicability.
Defect Detection in Industry
Because of the constraints of industrial production conditions, workpiece defects are difficult to avoid. Typical defects include debris, porosity, and cracks inside the workpiece.
These defects increase during the use of the workpiece and affect its performance. Therefore they cause the workpiece to fail, shortening its service life, and threatening the safety of the machine.
Image Fusion for defect identification in industry
The current defect detection algorithms are generally divided into two groups:
Agricultural Remote Sensing
Image fusion technology is also widely used in the field of agricultural remote sensing. By using agricultural remote sensing technology, farmers select the environment for the adaptation of plants and the detection of plant diseases.
Existing fusion technologies, including equipment such as ranging and optical detection, synthetic radar, and medium-resolution imaging spectrometers, all have applications in image fusion.
Image Fusion in Agricultural Remote Sensing
Researchers utilize a region-based fusion scheme for combining panchromatic, multispectral, and synthetic aperture radar images. In addition, some farmers combine spectral information, radar range data, and optical detection.
Advantages and Drawbacks of IF
Advantages of IF
Benefits of image fusion include:
Drawbacks of IF
Image fusion has certain limitations, such as:
Summary
Image fusion is an important technique for the integration, and evaluation of data from multiple sources (sensors). It has many applications in computer vision, medical imaging, and remote sensing.
Image fusions with complex nonlinear distortions contribute to the robustness of the most complex computer vision methods.
Related Products
Share this: