Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,508)

Search Parameters:
Keywords = RGB

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 9319 KiB  
Article
Blind Separation of Skin Chromophores from Multispectral Dermatological Images
by Mustapha Zokay and Hicham Saylani
Diagnostics 2024, 14(20), 2288; https://doi.org/10.3390/diagnostics14202288 - 14 Oct 2024
Abstract
Background/Objectives: Based on Blind Source Separation and the use of multispectral imaging, the new approach we propose in this paper aims to improve the estimation of the concentrations of the main skin chromophores (melanin, oxyhemoglobin and deoxyhemoglobin), while considering shading as a [...] Read more.
Background/Objectives: Based on Blind Source Separation and the use of multispectral imaging, the new approach we propose in this paper aims to improve the estimation of the concentrations of the main skin chromophores (melanin, oxyhemoglobin and deoxyhemoglobin), while considering shading as a fully-fledged source. Methods: In this paper, we demonstrate that the use of the Infra-Red spectral band, in addition to the traditional RGB spectral bands of dermatological images, allows us to model the image provided by each spectral band as a mixture of the concentrations of the three chromophores in addition to that of the shading, which are estimated through four steps using Blind Source Separation. Results: We studied the performance of our new method on a database of real multispectral dermatological images of melanoma by proposing a new quantitative performances measurement criterion based on mutual information. We then validated these performances on a database of multispectral dermatological images that we simulated using our own new protocol. Conclusions: All the results obtained demonstrated the effectiveness of our new approach for estimating the concentrations of the skin chromophores from a multispectral dermatological image, compared to traditional approaches that consist of using only the RGB image by neglecting shading. Full article
Show Figures

Figure 1

20 pages, 3542 KiB  
Article
Green Light Drives Embryonic Photosynthesis and Protein Accumulation in Cotyledons of Developing Pea (Pisum sativum L.) Seeds
by Nataliia Stepanova, Elena Tarakhovskaya, Alena Soboleva, Anastasia Orlova, Aditi Basnet, Anastasia Smolenskaya, Nadezhda Frolova, Tatiana Bilova, Anastasia Kamionskaya, Andrej Frolov, Sergei Medvedev and Galina Smolikova
Agronomy 2024, 14(10), 2367; https://doi.org/10.3390/agronomy14102367 - 14 Oct 2024
Viewed by 192
Abstract
Photosynthesis is a vital process for seed productivity. It occurs in the leaves and provides developing seeds with the necessary nutrients. Moreover, many crops require photochemical reactions inside the seeds for proper development. The present study aimed to investigate Pisum sativum L. seeds [...] Read more.
Photosynthesis is a vital process for seed productivity. It occurs in the leaves and provides developing seeds with the necessary nutrients. Moreover, many crops require photochemical reactions inside the seeds for proper development. The present study aimed to investigate Pisum sativum L. seeds at the middle stage of maturation, which is characterized by the active synthesis of nutrient reserves. Embryonic photosynthesis represents a crucial process to produce cells’ NADP(H) and ATP, which are necessary to convert sucrose into reserve biopolymers. However, it remains unclear how the pea embryo, covered by a coat and pericarp, receives sufficient light to provide energy for photochemical reactions. Recent studies have demonstrated that the photosynthetically active radiation reaching the developing pea embryo has a high proportion of green light. In addition, green light can be utilized in foliar photosynthesis by plants cultivated in shaded conditions. Here, we addressed the role of green light in seed development. Pea plants were cultivated under red and blue (RB) LEDs or red, green, and blue (RGB) LEDs. A Chl a fluorescence transient based on OJIP kinetics was detected at the periphery of the cotyledons isolated from developing seeds. Our findings showed that the addition of green light resulted in an increase in photochemical activity. Furthermore, the mature seeds that developed in the RGB module had a significantly higher weight and more storage proteins. Using a metabolomics approach, we also detected significant differences in the levels of organic acids, carbohydrates, nucleotide monophosphates, and nitrogenous substances between the RB and RGB conditions. Under RGB light, the cotyledons contained more ornithine, tryptophan, arginine, and aspartic acid. These changes indicate an impact of green light on the ornithine–urea cycle and polyamine biosynthesis. These results allow for a deeper understanding of the photochemical processes in embryos of developing seeds grown under a low light intensity. The photosynthetic system in the embryo cell adapts to the shade conditions by using green light. Full article
(This article belongs to the Special Issue Seeds: Chips of Agriculture)
Show Figures

Figure 1

21 pages, 12032 KiB  
Article
A Coffee Plant Counting Method Based on Dual-Channel NMS and YOLOv9 Leveraging UAV Multispectral Imaging
by Xiaorui Wang, Chao Zhang, Zhenping Qiang, Chang Liu, Xiaojun Wei and Fengyun Cheng
Remote Sens. 2024, 16(20), 3810; https://doi.org/10.3390/rs16203810 - 13 Oct 2024
Viewed by 373
Abstract
Accurate coffee plant counting is a crucial metric for yield estimation and a key component of precision agriculture. While multispectral UAV technology provides more accurate crop growth data, the varying spectral characteristics of coffee plants across different phenological stages complicate automatic plant counting. [...] Read more.
Accurate coffee plant counting is a crucial metric for yield estimation and a key component of precision agriculture. While multispectral UAV technology provides more accurate crop growth data, the varying spectral characteristics of coffee plants across different phenological stages complicate automatic plant counting. This study compared the performance of mainstream YOLO models for coffee detection and segmentation, identifying YOLOv9 as the best-performing model, with it achieving high precision in both detection (P = 89.3%, mAP50 = 94.6%) and segmentation performance (P = 88.9%, mAP50 = 94.8%). Furthermore, we studied various spectral combinations from UAV data and found that RGB was most effective during the flowering stage, while RGN (Red, Green, Near-infrared) was more suitable for non-flowering periods. Based on these findings, we proposed an innovative dual-channel non-maximum suppression method (dual-channel NMS), which merges YOLOv9 detection results from both RGB and RGN data, leveraging the strengths of each spectral combination to enhance detection accuracy and achieving a final counting accuracy of 98.4%. This study highlights the importance of integrating UAV multispectral technology with deep learning for coffee detection and offers new insights for the implementation of precision agriculture. Full article
Show Figures

Figure 1

24 pages, 13862 KiB  
Article
Depth Video-Based Secondary Action Recognition in Vehicles via Convolutional Neural Network and Bidirectional Long Short-Term Memory with Spatial Enhanced Attention Mechanism
by Weirong Shao, Mondher Bouazizi and Ohtuski Tomoaki
Sensors 2024, 24(20), 6604; https://doi.org/10.3390/s24206604 (registering DOI) - 13 Oct 2024
Viewed by 366
Abstract
Secondary actions in vehicles are activities that drivers engage in while driving that are not directly related to the primary task of operating the vehicle. Secondary Action Recognition (SAR) in drivers is vital for enhancing road safety and minimizing accidents related to distracted [...] Read more.
Secondary actions in vehicles are activities that drivers engage in while driving that are not directly related to the primary task of operating the vehicle. Secondary Action Recognition (SAR) in drivers is vital for enhancing road safety and minimizing accidents related to distracted driving. It also plays an important part in modern car driving systems such as Advanced Driving Assistance Systems (ADASs), as it helps identify distractions and predict the driver’s intent. Traditional methods of action recognition in vehicles mostly rely on RGB videos, which can be significantly impacted by external conditions such as low light levels. In this research, we introduce a novel method for SAR. Our approach utilizes depth-video data obtained from a depth sensor located in a vehicle. Our methodology leverages the Convolutional Neural Network (CNN), which is enhanced by the Spatial Enhanced Attention Mechanism (SEAM) and combined with Bidirectional Long Short-Term Memory (Bi-LSTM) networks. This method significantly enhances action recognition ability in depth videos by improving both the spatial and temporal aspects. We conduct experiments using K-fold cross validation, and the experimental results show that on the public benchmark dataset Drive&Act, our proposed method shows significant improvement in SAR compared to the state-of-the-art methods, reaching an accuracy of about 84% in SAR in depth videos. Full article
(This article belongs to the Special Issue Multi-Sensor Systems for Object Tracking—2nd Edition)
Show Figures

Figure 1

15 pages, 3259 KiB  
Article
Enhancing the Performance of Unmanned Aerial Vehicle-Based Estimation of Rape Chlorophyll Content by Reducing the Impact of Crop Coverage
by Yaxiao Niu, Longfei Xu, Yanni Zhang, Lizhang Xu, Qingzhen Zhu, Aichen Wang, Shenjin Huang and Liyuan Zhang
Drones 2024, 8(10), 578; https://doi.org/10.3390/drones8100578 - 12 Oct 2024
Viewed by 319
Abstract
Estimating leaf chlorophyll content (LCC) in a timely manner and accurately is of great significance for the precision management of rape. The spectral index derived from UAV images has been adopted as a non-destructive and efficient way to map LCC. However, soil background [...] Read more.
Estimating leaf chlorophyll content (LCC) in a timely manner and accurately is of great significance for the precision management of rape. The spectral index derived from UAV images has been adopted as a non-destructive and efficient way to map LCC. However, soil background impairs the performance of UAV-based LCC estimation, limiting the accuracy and applicability of the LCC estimation model, and this issue remains to be addressed. Thus, this research was conducted to study the influence of soil pixels in UAV RGB images on LCC estimation. UAV campaigns were conducted from overwintering to flowering stages to cover the process of soil background being gradually covered by rapeseed plants. Three planting densities of 11.25, 18.75, and 26.26 g/m2 were chosen to further enrich the different soil background percentage levels, namely, the rape fractional vegetation coverage (FVC) levels. The results showed that, compared to the insignificant difference observed for the ground measured LCC at a certain growth stage, a significant difference was found for most of the spectral indices extracted without soil background removal, indicating the influence of soil background. Removing soil background during the extraction of the spectral index enhanced the LCC estimation accuracy, with the coefficient of determination (R2) increasing from 0.58 to 0.68 and the root mean square error (RMSE) decreasing from 5.19 to 4.49. At the same time, the applicability of the LCC estimation model for different plant densities (FVC levels) was also enhanced. The lower the planting density, the greater the enhancement. R2 increased from 0.53 to 0.70, and the RMSE decreased from 5.30 to 4.81 under a low planting density of 11.25 g/m2. These findings indicate that soil background removal significantly enhances the performance of UAV-based rape LCC estimation, particularly under various FVC conditions. Full article
(This article belongs to the Special Issue UAS in Smart Agriculture: 2nd Edition)
Show Figures

Figure 1

28 pages, 7076 KiB  
Article
Coupling Image-Fusion Techniques with Machine Learning to Enhance Dynamic Monitoring of Nitrogen Content in Winter Wheat from UAV Multi-Source
by Xinwei Li, Xiangxiang Su, Jun Li, Sumera Anwar, Xueqing Zhu, Qiang Ma, Wenhui Wang and Jikai Liu
Agriculture 2024, 14(10), 1797; https://doi.org/10.3390/agriculture14101797 - 12 Oct 2024
Viewed by 364
Abstract
Plant nitrogen concentration (PNC) is a key indicator reflecting the growth and development status of plants. The timely and accurate monitoring of plant PNC is of great significance for the refined management of crop nutrition in the field. The rapidly developing sensor technology [...] Read more.
Plant nitrogen concentration (PNC) is a key indicator reflecting the growth and development status of plants. The timely and accurate monitoring of plant PNC is of great significance for the refined management of crop nutrition in the field. The rapidly developing sensor technology provides a powerful means for monitoring crop PNC. Although RGB images have rich spatial information, they lack the spectral information of the red edge and near infrared bands, which are more sensitive to vegetation. Conversely, multispectral images offer superior spectral resolution but typically lag in spatial detail compared to RGB images. Therefore, the purpose of this study is to improve the accuracy and efficiency of crop PNC monitoring by combining the advantages of RGB images and multispectral images through image-fusion technology. This study was based on the booting, heading, and early-filling stages of winter wheat, synchronously acquiring UAV RGB and MS data, using Gram–Schmidt (GS) and principal component (PC) image-fusion methods to generate fused images and evaluate them with multiple image-quality indicators. Subsequently, models for predicting wheat PNC were constructed using machine-selection algorithms such as RF, GPR, and XGB. The results show that the RGB_B1 image contains richer image information and more image details compared to other bands. The GS image-fusion method is superior to the PC method, and the performance of fusing high-resolution RGB_B1 band images with MS images using the GS method is optimal. After image fusion, the correlation between vegetation indices (VIs) and wheat PNC has been enhanced to varying degrees in different growth periods, significantly enhancing the response ability of spectral information to wheat PNC. To comprehensively assess the potential of fused images in estimating wheat PNC, this study fully compared the performance of PNC models before and after fusion using machine learning algorithms such as Random Forest (RF), Gaussian Process Regression (GPR), and eXtreme Gradient Boosting (XGB). The results show that the model established by the fusion image has high stability and accuracy in a single growth period, multiple growth periods, different varieties, and different nitrogen treatments, making it significantly better than the MS image. The most significant enhancements were during the booting to early-filling stages, particularly with the RF algorithm, which achieved an 18.8% increase in R2, a 26.5% increase in RPD, and a 19.7% decrease in RMSE. This study provides an effective technical means for the dynamic monitoring of crop nutritional status and provides strong technical support for the precise management of crop nutrition. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

14 pages, 6262 KiB  
Article
Degradation-Guided Multi-Modal Fusion Network for Depth Map Super-Resolution
by Lu Han, Xinghu Wang, Fuhui Zhou and Diansheng Wu
Electronics 2024, 13(20), 4020; https://doi.org/10.3390/electronics13204020 - 12 Oct 2024
Viewed by 195
Abstract
Depth map super-resolution (DSR) is a technique aimed at restoring high-resolution (HR) depth maps from low-resolution (LR) depth maps. In this process, color images are commonly used as guidance to enhance the restoration procedure. However, the intricate degradation of LR depth poses a [...] Read more.
Depth map super-resolution (DSR) is a technique aimed at restoring high-resolution (HR) depth maps from low-resolution (LR) depth maps. In this process, color images are commonly used as guidance to enhance the restoration procedure. However, the intricate degradation of LR depth poses a challenge, and previous image-guided DSR approaches, which implicitly model the degradation in the spatial domain, often fall short of producing satisfactory results. To address this challenge, we propose a novel approach called the Degradation-Guided Multi-modal Fusion Network (DMFNet). DMFNet explicitly characterizes the degradation and incorporates multi-modal fusion in both spatial and frequency domains to improve the depth quality. Specifically, we first introduce the deep degradation regularization loss function, which enables the model to learn the explicit degradation from the LR depth maps. Simultaneously, DMFNet converts the color images and depth maps into spectrum representations to provide comprehensive multi-domain guidance. Consequently, we present the multi-modal fusion block to restore the depth maps by leveraging both the RGB-D spectrum representations and the depth degradation. Extensive experiments demonstrate that DMFNet achieves state-of-the-art (SoTA) performance on four benchmarks, namely the NYU-v2, Middlebury, Lu, and RGB-D-D datasets. Full article
(This article belongs to the Special Issue Advances in Data-Driven Artificial Intelligence)
Show Figures

Figure 1

20 pages, 6262 KiB  
Article
YPR-SLAM: A SLAM System Combining Object Detection and Geometric Constraints for Dynamic Scenes
by Xukang Kan, Gefei Shi, Xuerong Yang and Xinwei Hu
Sensors 2024, 24(20), 6576; https://doi.org/10.3390/s24206576 (registering DOI) - 12 Oct 2024
Viewed by 174
Abstract
Traditional SLAM systems assume a static environment, but moving objects break this ideal assumption. In the real world, moving objects can greatly influence the precision of image matching and camera pose estimation. In order to solve these problems, the YPR-SLAM system is proposed. [...] Read more.
Traditional SLAM systems assume a static environment, but moving objects break this ideal assumption. In the real world, moving objects can greatly influence the precision of image matching and camera pose estimation. In order to solve these problems, the YPR-SLAM system is proposed. First of all, the system includes a lightweight YOLOv5 detection network for detecting both dynamic and static objects, which provides pre-dynamic object information to the SLAM system. Secondly, utilizing the prior information of dynamic targets and the depth image, a method of geometric constraint for removing motion feature points from the depth image is proposed. The Depth-PROSAC algorithm is used to differentiate the dynamic and static feature points so that dynamic feature points can be removed. At last, the dense cloud map is constructed by the static feature points. The YPR-SLAM system is an efficient combination of object detection and geometry constraint in a tightly coupled way, eliminating motion feature points and minimizing their adverse effects on SLAM systems. The performance of the YPR-SLAM was assessed on the public TUM RGB-D dataset, and it was found that YPR-SLAM was suitable for dynamic situations. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

15 pages, 3752 KiB  
Article
Camellia oleifera Tree Detection and Counting Based on UAV RGB Image and YOLOv8
by Renxu Yang, Debao Yuan, Maochen Zhao, Zhao Zhao, Liuya Zhang, Yuqing Fan, Guangyu Liang and Yifei Zhou
Agriculture 2024, 14(10), 1789; https://doi.org/10.3390/agriculture14101789 - 12 Oct 2024
Viewed by 306
Abstract
The detection and counting of Camellia oleifera trees are important parts of the yield estimation of Camellia oleifera. The ability to identify and count Camellia oleifera trees quickly has always been important in the context of research on the yield estimation of [...] Read more.
The detection and counting of Camellia oleifera trees are important parts of the yield estimation of Camellia oleifera. The ability to identify and count Camellia oleifera trees quickly has always been important in the context of research on the yield estimation of Camellia oleifera. Because of their specific growing environment, it is a difficult task to identify and count Camellia oleifera trees with high efficiency. In this paper, based on a UAV RGB image, three different types of datasets, i.e., a DOM dataset, an original image dataset, and a cropped original image dataset, were designed. Combined with the YOLOv8 model, the detection and counting of Camellia oleifera trees were carried out. By comparing YOLOv9 and YOLOv10 in four evaluation indexes, including precision, recall, mAP, and F1 score, Camellia oleifera trees in two areas were selected for prediction and compared with the real values. The experimental results show that the cropped original image dataset was better for the recognition and counting of Camellia oleifera, and the mAP values were 8% and 11% higher than those of the DOM dataset and the original image dataset, respectively. Compared to YOLOv5, YOLOv7, YOLOv9, and YOLOv10, YOLOv8 performed better in terms of the accuracy and recall rate, and the mAP improved by 3–8%, reaching 0.82. Regression analysis was performed on the predicted and measured values, and the average R2 reached 0.94. This research shows that a UAV RGB image combined with YOLOv8 provides an effective solution for the detection and counting of Camellia oleifera trees, which is of great significance for Camellia oleifera yield estimation and orchard management. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

26 pages, 7527 KiB  
Article
Ultrasonic Weld Quality Inspection Involving Strength Prediction and Defect Detection in Data-Constrained Training Environments
by Reenu Mohandas, Patrick Mongan and Martin Hayes
Sensors 2024, 24(20), 6553; https://doi.org/10.3390/s24206553 - 11 Oct 2024
Viewed by 485
Abstract
Welding is an extensively used technique in manufacturing, and as for every other process, there is the potential for defects in the weld joint that could be catastrophic to the manufactured products. Different welding processes use different parameter settings, which greatly impact the [...] Read more.
Welding is an extensively used technique in manufacturing, and as for every other process, there is the potential for defects in the weld joint that could be catastrophic to the manufactured products. Different welding processes use different parameter settings, which greatly impact the quality of the final welded products. The focus of research in weld defect detection is to develop a non-destructive testing method for weld quality assessment based on observing the weld with an RGB camera. Deep learning techniques have been widely used in the domain of weld defect detection in recent times, but the majority of them use, for example, X-ray images. An RGB image-based solution is attractive, as RGB cameras are comparatively inexpensive compared to X-ray image solutions. However, the number of publicly available RGB image datasets for weld defect detection is comparatively lower than that of X-ray image datasets. This work achieves a complete weld quality assessment involving lap shear strength prediction and visual weld defect detection from an extremely limited dataset. First, a multimodal dataset is generated by the fusion of image data features extracted using a convolutional autoencoder (CAE) designed in this experiment and input parameter settings data. The fusion of the dataset reduced lap shear strength (LSS) prediction errors by 34% compared to prediction errors using only input parameter settings data. This is a promising result, considering the extremely small dataset size. This work also achieves visual weld defect detection on the same limited dataset with the help of an ultrasonic weld defect dataset generated using offline and online data augmentation. The weld defect detection achieves an accuracy of 74%, again a promising result that meets standard requirements. The combination of lap shear strength prediction and visual defect detection leads to a complete inspection to avoid premature failure of the ultrasonic weld joints. The weld defect detection was compared against the publicly available image dataset for surface defect detection. Full article
(This article belongs to the Special Issue Deep-Learning-Based Defect Detection for Smart Manufacturing)
Show Figures

Figure 1

15 pages, 6871 KiB  
Article
A Trianalyte µPAD for Simultaneous Determination of Iron, Zinc, and Manganese Ions
by Barbara Rozbicka, Robert Koncki and Marta Fiedoruk-Pogrebniak
Molecules 2024, 29(20), 4805; https://doi.org/10.3390/molecules29204805 - 11 Oct 2024
Viewed by 188
Abstract
In this work, a microfluidic paper-based analytical device (µPAD) for simultaneous detection of Fe, Zn, and Mn ions using immobilized chromogenic reagents Ferene S, xylenol orange, and 1-(2-pyridylazo)-2-naphthol, respectively, is presented. As the effective recognition of analytes via respective chromogens takes place under [...] Read more.
In this work, a microfluidic paper-based analytical device (µPAD) for simultaneous detection of Fe, Zn, and Mn ions using immobilized chromogenic reagents Ferene S, xylenol orange, and 1-(2-pyridylazo)-2-naphthol, respectively, is presented. As the effective recognition of analytes via respective chromogens takes place under extremely different pH conditions, experiments reported in this publication are focused on optimization of the µPAD architecture allowing for the elimination of potential cross effects. The paper-based microfluidic device was fabricated using low-cost and well-reproducible wax-printing technology. For optical detection of color changes, an ordinary office scanner and self-made RGB-data processing program were applied. Optimized and stable over time, µPADs allow fast, selective, and reproducible multianalyte determinations at submillimolar levels of respective heavy metal ions, which was confirmed by results of the analysis of solutions mimicking real samples of wastewater. The presented concept of simultaneous determination of different analytes that required extremely different conditions for detection can be useful for the development of other multianalyte microfluidic paper-based devices in the µPAD format. Full article
Show Figures

Figure 1

13 pages, 7413 KiB  
Article
A Study on Enhancing the Visual Fidelity of Aviation Simulators Using WGAN-GP for Remote Sensing Image Color Correction
by Chanho Lee, Hyukjin Kwon, Hanseon Choi, Jonggeun Choi, Ilkyun Lee, Byungkyoo Kim, Jisoo Jang and Dongkyoo Shin
Appl. Sci. 2024, 14(20), 9227; https://doi.org/10.3390/app14209227 - 11 Oct 2024
Viewed by 338
Abstract
When implementing outside-the-window (OTW) visuals in aviation tactical simulators, maintaining terrain image color consistency is critical for enhancing pilot immersion and focus. However, due to various environmental factors, inconsistent image colors in terrain can cause visual confusion and diminish realism. To address these [...] Read more.
When implementing outside-the-window (OTW) visuals in aviation tactical simulators, maintaining terrain image color consistency is critical for enhancing pilot immersion and focus. However, due to various environmental factors, inconsistent image colors in terrain can cause visual confusion and diminish realism. To address these issues, a color correction technique based on a Wasserstein Generative Adversarial Network with Gradient Penalty (WGAN-GP) is proposed. The proposed WGAN-GP model utilizes multi-scale feature extraction and Wasserstein distance to effectively measure and adjust the color distribution difference between the input image and the reference image. This approach can preserve the texture and structural characteristics of the image while maintaining color consistency. In particular, by converting Bands 2, 3, and 4 of the BigEarthNet-S2 dataset into RGB images as the reference image and preprocessing the reference image to serve as the input image, it is demonstrated that the proposed WGAN-GP model can handle large-scale remote sensing images containing various lighting conditions and color differences. The experimental results showed that the proposed WGAN-GP model outperformed traditional methods, such as histogram matching and color transfer, and was effective in reflecting the style of the reference image to the target image while maintaining the structural elements of the target image during the training process. Quantitative analysis demonstrated that the mid-stage model achieved a PSNR of 28.93 dB and an SSIM of 0.7116, which significantly outperforms traditional methods. Furthermore, the LPIPS score was reduced to 0.3978, indicating improved perceptual similarity. This approach can contribute to improving the visual elements of the simulator to enhance pilot immersion and has the potential to significantly reduce time and costs compared to the manual methods currently used by the Republic of Korea Air Force. Full article
(This article belongs to the Special Issue Applications of Machine Learning Algorithms in Remote Sensing)
Show Figures

Figure 1

18 pages, 7228 KiB  
Article
RBFNN-PSO Intelligent Synchronisation Method for Sprott B Chaotic Systems with External Noise and Its Application in an Image Encryption System
by Yanpeng Zhang, Jian Zeng, Wenhao Yan and Qun Ding
Entropy 2024, 26(10), 855; https://doi.org/10.3390/e26100855 - 10 Oct 2024
Viewed by 236
Abstract
In the past two decades, research in the field of chaotic synchronization has attracted extensive attention from scholars, and at the same time, more synchronization methods, such as chaotic master–slave synchronization, projection synchronization, sliding film synchronization, fractional-order synchronization and so on, have been [...] Read more.
In the past two decades, research in the field of chaotic synchronization has attracted extensive attention from scholars, and at the same time, more synchronization methods, such as chaotic master–slave synchronization, projection synchronization, sliding film synchronization, fractional-order synchronization and so on, have been proposed and applied to chaotic secure communication. In this paper, based on radial basis function neural network theory and the particle swarm optimisation algorithm, the RBFNN-PSO synchronisation method is proposed for the Sprott B chaotic system with external noise. The RBFNN controller is constructed, and its parameters are used as the particle swarm particle optimisation parameters, and the optimal values of the controller parameters are obtained by the PSO training method, which overcomes the influence of external noise and achieves the synchronisation of the master–slave system. Then, it is shown by numerical simulation and analysis that the scheme has a good performance against external noise. Because the Sprott B system has multiple chaotic attractors with richer dynamics, the synchronization system based on Sprott B chaos is applied to the image encryption system. In particular, the Zigzag disambiguation method for top corner rotation and RGB channel selection is proposed, and the master–slave chaotic system synchronisation sequences are diffused to the disambiguated data streams, respectively. Therefore, the encryption and decryption of image transmission are implemented and the numerical simulation results are given, the random distribution characteristics of encrypted images are analysed using histogram and Shannon entropy methods, and the final results achieve the expected results. Full article
Show Figures

Figure 1

19 pages, 8953 KiB  
Article
Leveraging Multimodal Large Language Models (MLLMs) for Enhanced Object Detection and Scene Understanding in Thermal Images for Autonomous Driving Systems
by Huthaifa I. Ashqar, Taqwa I. Alhadidi, Mohammed Elhenawy and Nour O. Khanfar
Automation 2024, 5(4), 508-526; https://doi.org/10.3390/automation5040029 - 10 Oct 2024
Viewed by 684
Abstract
The integration of thermal imaging data with multimodal large language models (MLLMs) offers promising advancements for enhancing the safety and functionality of autonomous driving systems (ADS) and intelligent transportation systems (ITS). This study investigates the potential of MLLMs, specifically GPT-4 Vision Preview and [...] Read more.
The integration of thermal imaging data with multimodal large language models (MLLMs) offers promising advancements for enhancing the safety and functionality of autonomous driving systems (ADS) and intelligent transportation systems (ITS). This study investigates the potential of MLLMs, specifically GPT-4 Vision Preview and Gemini 1.0 Pro Vision, for interpreting thermal images for applications in ADS and ITS. Two primary research questions are addressed: the capacity of these models to detect and enumerate objects within thermal images, and to determine whether pairs of image sources represent the same scene. Furthermore, we propose a framework for object detection and classification by integrating infrared (IR) and RGB images of the same scene without requiring localization data. This framework is particularly valuable for enhancing the detection and classification accuracy in environments where both IR and RGB cameras are essential. By employing zero-shot in-context learning for object detection and the chain-of-thought technique for scene discernment, this study demonstrates that MLLMs can recognize objects such as vehicles and individuals with promising results, even in the challenging domain of thermal imaging. The results indicate a high true positive rate for larger objects and moderate success in scene discernment, with a recall of 0.91 and a precision of 0.79 for similar scenes. The integration of IR and RGB images further enhances detection capabilities, achieving an average precision of 0.93 and an average recall of 0.56. This approach leverages the complementary strengths of each modality to compensate for individual limitations. This study highlights the potential of combining advanced AI methodologies with thermal imaging to enhance the accuracy and reliability of ADS, while identifying areas for improvement in model performance. Full article
Show Figures

Figure 1

16 pages, 11255 KiB  
Article
Cherry Tomato Detection for Harvesting Using Multimodal Perception and an Improved YOLOv7-Tiny Neural Network
by Yingqi Cai, Bo Cui, Hong Deng, Zhi Zeng, Qicong Wang, Dajiang Lu, Yukang Cui and Yibin Tian
Agronomy 2024, 14(10), 2320; https://doi.org/10.3390/agronomy14102320 - 9 Oct 2024
Viewed by 473
Abstract
Robotic fruit harvesting has great potential to revolutionize agriculture, but detecting cherry tomatoes in farming environments still faces challenges in accuracy and efficiency. To overcome the shortcomings of existing cherry tomato detection methods for harvesting, this study introduces a deep-learning-based cherry tomato detection [...] Read more.
Robotic fruit harvesting has great potential to revolutionize agriculture, but detecting cherry tomatoes in farming environments still faces challenges in accuracy and efficiency. To overcome the shortcomings of existing cherry tomato detection methods for harvesting, this study introduces a deep-learning-based cherry tomato detection scheme for robotic harvesting in greenhouses using multimodal RGB-D perception and an improved YOLOv7-tiny Cherry Tomato Detection (YOLOv7-tiny-CTD) network, which has been modified from the original YOLOv7-tiny by eliminating the “Objectness” output layer, introducing a new “Classness” method for the prediction box, and incorporating a new hybrid non-maximum suppression. Acquired RGB-D images undergo preprocessing such as color space transformation, point cloud normal vector angle computation, and multimodal regions of interest segmentation before being fed into the YOLOv7-tiny-CTD. The proposed method was tested using an AGV-based robot in a greenhouse cherry tomato farming facility. The results indicate that the multimodal perception and deep learning method improves detection precision and accuracy over existing methods while running in real time, and the robot achieved over 80% successful picking rates in two-trial mode in the greenhouse farm, showing promising potential for practical harvesting applications. Full article
(This article belongs to the Special Issue AI, Sensors and Robotics for Smart Agriculture—2nd Edition)
Show Figures

Figure 1

Back to TopTop