Advanced Image Stitching Method for Dual-Sensor Inspection
Abstract
:1. Introduction
- We propose a self-supervised auto-encoder feature detection technique for enhancing the quality and quantity of feature points in infrared and visible images;
- We also employ a powerful feature matching algorithm based on graph neural networks to identify and remove the mismatched features robustly;
- Lastly, we develop perspective-distortion-free image stitching software for dual-sensor inspection especially for low-texture conditions.
2. Literature Review
2.1. Feature Detection and Descriptor
2.2. Feature Matching
Image Stitching
3. Materials and Methods
Algorithm 1: Advance Image Stitching Algorithm for Dual-Sensor Inspection |
3.1. Feature Detection and Description Phase
Self-Training Auto-Encoder for Unified Dual-Sensor Feature Detection and Description
3.2. Feature Matching Phase
- Translation: Shifting the position of one image relative to another;
- Rotation: Rotating an image to align features;
- Scaling: Adjusting the size of an image to match the scale of the reference image;
- Homography: A more general transformation that includes translation, rotation, scaling, and skewing. It is particularly useful for correcting perspective distortions.
- Perspective Distortion Correction: Geometric transformations, such as homography, can correct for perspective distortions when objects in the scene are viewed from different angles. It is particularly relevant when capturing images with wide-angle lenses or from non-ideal shooting positions;
- Seamless Alignment: Applying appropriate transformations ensures that key features in the overlapping regions of adjacent images align correctly. This alignment is critical for creating a visually coherent and distortion-free stitched image;
- Global Adjustment: Geometric transformations allow for global adjustments, ensuring that the entire set of images contributes cohesively to the stitched result. This is essential for avoiding artifacts and maintaining a natural appearance.
3.3. Multi-Image Stitching Phase
4. Results and Discussion
4.1. Dataset and Implementation Details
4.2. Results of Feature Detection and Description Phase
4.3. Results of Feature Matching Phase
- Difficulty in handling repetitive patterns: Traditional methods may struggle to distinguish between repeated patterns, leading to ambiguous matches;
- Repeatability: Changes in lighting conditions and surface reflectance can result in inconsistent feature points and matches;
- The proposed method specializes in feature detection and description methods appropriate for complex structures which can adapt to these unique conditions and provide more reliable infrastructure inspection and assessment results.
- Sparse Keypoint Distributions: The analysis of ORB and SIFT feature points reveals a sparse distribution, limiting the possibilities for feature matching. This scarcity can hinder the overall effectiveness of these methods;
- Inaccurate Matches: The inherent limitations of ORB, SIFT, and AKAZE become apparent in scenarios involving parallax or poor texture, where keypoints may fail to align accurately between images. This discrepancy results in incorrect feature matches, adversely affecting the reliability of the matching process;
- Difficulty in Handling Repetitive Patterns: Traditional methods, including ORB and SIFT, face challenges in distinguishing between repeated patterns. This difficulty leads to ambiguous matches, introducing uncertainty into the feature matching outcomes;
- Repeatability Challenges: Changes in lighting conditions and surface reflectance pose challenges to the repeatability of feature points and matches in ORB, SIFT, and AKAZE. This inconsistency in performance can compromise the reliability of these methods in real-world applications.
4.4. Results of Image Stitching Phase
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
IRT | Infrared thermography |
NDE | Non-Destructive Evaluation |
NDT | Non-destructive testing |
CNN | Convolutional neural network |
AKAZE | Accelerated Keypoint and Zernike Moment Descriptor |
ORB | Oriented FAST and Rotated BRIEF |
SIFT | Scale-Invariant Feature Transform |
RANSAC | Random Sample Consensus |
References
- Kot, P.; Muradov, M.; Gkantou, M.; Kamaris, G.S.; Hashim, K.; Yeboah, D. Recent advancements in non-destructive testing techniques for structural health monitoring. Appl. Sci. 2021, 11, 2750. [Google Scholar] [CrossRef]
- Fuentes, S.; Tongson, E.; Gonzalez Viejo, C. Urban green infrastructure monitoring using remote sensing from integrated visible and thermal infrared cameras mounted on a moving vehicle. Sensors 2021, 21, 295. [Google Scholar] [CrossRef] [PubMed]
- Chen, J.; Li, Z.; Peng, C.; Wang, Y.; Gong, W. UAV Image Stitching Based on Optimal Seam and Half-Projective Warp. Remote Sens. 2022, 14, 1068. [Google Scholar] [CrossRef]
- Shahsavarani, S.; Lopez, F.; Maldague, X. Multi-modal image processing pipeline for NDE of structures and industrial assets. In Thermosense: Thermal Infrared Applications XLV; SPIE: Bellingham, WA, USA, 2023; Volume 12536, pp. 255–264. [Google Scholar]
- Ghiass, R.S.; Arandjelović, O.; Bendada, H.; Maldague, X. A unified framework for thermal face recognition. In Neural Information Processing, Proceedings of the 21st International Conference, ICONIP 2014, Kuching, Malaysia, 3–6 November 2014; Proceedings, Part II; Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 335–343. [Google Scholar]
- Ghiass, R.S.; Arandjelović, O.; Bendada, H.; Maldague, X. Illumination-invariant face recognition from a single image across extreme pose using a dual dimension AAM ensemble in the thermal infrared spectrum. In Proceedings of the 2013 International Joint Conference on Neural Networks (IJCNN), Dallas, TX, USA, 4–9 August 2013; pp. 1–10. [Google Scholar]
- Nooralishahi, P.; Ramos, G.; Pozzer, S.; Ibarra-Castanedo, C.; Lopez, F.; Maldague, X.P. Texture analysis to enhance drone-based multi-modal inspection of structures. Drones 2022, 6, 407. [Google Scholar] [CrossRef]
- Yi, K.M.; Trulls, E.; Lepetit, V.; Fua, P. Lift: Learned invariant feature transform. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 467–483. [Google Scholar]
- Zhang, L.; Rusinkiewicz, S. Learning to detect features in texture images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6325–6333. [Google Scholar]
- DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar]
- Lenc, K.; Vedaldi, A. Learning covariant feature detectors. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–10 October 2016; pp. 100–117. [Google Scholar]
- Savinov, N.; Seki, A.; Ladicky, L.; Sattler, T.; Pollefeys, M. Quad-networks: Unsupervised learning to rank for interest point detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1822–1830. [Google Scholar]
- Ono, Y.; Trulls, E.; Fua, P.; Yi, K.M. LF-NET: Learning local features from images. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; Volume 31, pp. 6234–6244. [Google Scholar]
- Georgakis, G.; Karanam, S.; Wu, Z.; Ernst, J.; Kosecká, J. End-to-end learning of keypoint detector and descriptor for pose invariant 3D matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1965–1973. [Google Scholar]
- Barroso-Laguna, A.; Riba, E.; Ponsa, D.; Mikolajczyk, K. Key.net: Keypoint detection by handcrafted and learned CNN filters. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5836–5844. [Google Scholar]
- Verdie, Y.; Yi, K.; Fua, P.; Lepetit, V. Tilde: A temporally invariant learned detector. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 5279–5288. [Google Scholar]
- Zitova, B.; Flusser, J. Image registration methods: A survey. Image Vis. Comput. 2003, 21, 977–1000. [Google Scholar] [CrossRef]
- Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed]
- Dawn, S.; Saxena, V.; Sharma, B. Remote sensing image registration techniques: A survey. In Proceedings of the International Conference on Image and Signal Processing, Quebec City, QC, Canada, 30 June–2 July 2010; pp. 103–112. [Google Scholar]
- Sotiras, A.; Davatzikos, C.; Paragios, N. Deformable medical image registration: A survey. IEEE Trans. Med. Imaging 2013, 32, 1153. [Google Scholar] [CrossRef] [PubMed]
- Ferrante, E.; Paragios, N. Slice-to-volume medical image registration: A survey. Med. Image Anal. 2017, 39, 101–123. [Google Scholar] [CrossRef] [PubMed]
- Lawler, E.L. The quadratic assignment problem. Manag. Sci. 1963, 9, 586–599. [Google Scholar] [CrossRef]
- Belongie, S.; Malik, J.; Puzicha, J. Shape context: A new descriptor for shape matching and object recognition. In Proceedings of the Advances in Neural Information Processing Systems, Denver, CO, USA, 1 January 2000; pp. 831–837. [Google Scholar]
- Loiola, E.M.; de Abreu, N.M.M.; Boaventura-Netto, P.O.; Hahn, P.; Querido, T. A survey for the quadratic assignment problem. Eur. J. Oper. Res. 2007, 176, 657–690. [Google Scholar] [CrossRef]
- Ghosh, D.; Kaabouch, N.; Hu, W.C. A robust iterative super-resolution mosaicking algorithm using an adaptive and directional Huber-Markov regularization. J. Vis. Commun. Image Represent. 2016, 40, 98–110. [Google Scholar] [CrossRef]
- Brown, M.; Lowe, D.G. Automatic panoramic image stitching using invariant features. Int. J. Comput. Vis. 2007, 74, 59–73. [Google Scholar] [CrossRef]
- Lowe, D. Distinctive Image Features from Scale-Invariant Keypoints. Int. J. Comput. Vis. 2004, 60, 91–110. [Google Scholar] [CrossRef]
- Fischler, M.A.; Bolles, R.C. Random sample consensus: A paradigm for model fitting with applications to image analysis and automated cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
- Lin, W.Y.; Liu, S.; Matsushita, Y.; Ng, T.T.; Cheong, L.F. Smoothly varying affine stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Colorado Springs, CO, USA, 20–25 June 2011; pp. 345–352. [Google Scholar]
- Hafeez, J.; Lee, J.; Kwon, S.; Ha, S.; Hur, G.; Lee, S. Evaluating feature extraction methods with synthetic noise patterns for image-based modelling of texture-less objects. Remote Sens. 2020, 12, 3886. [Google Scholar] [CrossRef]
- Sun, J.; Shen, Z.; Wang, Y.; Bao, H.; Zhou, X. LoFTR: Detector-free local feature matching with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8922–8931. [Google Scholar]
- Wang, Y.; Yao, H.; Zhao, S. Auto-encoder based dimensionality reduction. Neurocomputing 2016, 184, 232–242. [Google Scholar] [CrossRef]
- Tschannen, M.; Bachem, O.; Lucic, M. Recent advances in autoencoder-based representation learning. arXiv 2018, arXiv:1812.05069. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
- Liu, W.; Wen, Y.; Yu, Z.; Yang, M. Large-Margin Softmax Loss for Convolutional Neural Networks. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 507–516. [Google Scholar]
- Neubeck, A.; Van Gool, L. Efficient Non-Maximum Suppression. In Proceedings of the 18th International Conference on Pattern Recognition (ICPR’06), Hong Kong, China, 20–24 August 2006; Volume 3, pp. 850–855. [Google Scholar]
- Cao, S.Y.; Hu, J.; Sheng, Z.; Shen, H.L. Iterative Deep Homography Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 1879–1888. [Google Scholar]
- Lin, T.Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. In Proceedings of the Computer Vision—ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings. Springer International Publishing: Berlin/Heidelberg, Germany, 2014; pp. 740–755. [Google Scholar]
- Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar]
- Rublee, E.; Rabaud, V.; Konolige, K.; Bradski, G. ORB: An efficient alternative to SIFT or SURF. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 2564–2571. [Google Scholar]
- Alcantarilla, P.F.; Bartoli, A.; Davison, A.J. KAZE features. In Proceedings of the Computer Vision—ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; pp. 214–227. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Shahsavarani, S.; Lopez, F.; Ibarra-Castanedo, C.; Maldague, X.P.V. Advanced Image Stitching Method for Dual-Sensor Inspection. Sensors 2024, 24, 3778. https://doi.org/10.3390/s24123778
Shahsavarani S, Lopez F, Ibarra-Castanedo C, Maldague XPV. Advanced Image Stitching Method for Dual-Sensor Inspection. Sensors. 2024; 24(12):3778. https://doi.org/10.3390/s24123778
Chicago/Turabian StyleShahsavarani, Sara, Fernando Lopez, Clemente Ibarra-Castanedo, and Xavier P. V. Maldague. 2024. "Advanced Image Stitching Method for Dual-Sensor Inspection" Sensors 24, no. 12: 3778. https://doi.org/10.3390/s24123778