Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,373)

Search Parameters:
Keywords = YOLOv5s

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 10212 KiB  
Article
YOLOv9s-Pear: A Lightweight YOLOv9s-Based Improved Model for Young Red Pear Small-Target Recognition
by Yi Shi, Zhen Duan, Shunhao Qing, Long Zhao, Fei Wang and Xingcan Yuwen
Agronomy 2024, 14(9), 2086; https://doi.org/10.3390/agronomy14092086 (registering DOI) - 12 Sep 2024
Abstract
With the advancement of computer vision technology, the demand for fruit recognition in agricultural automation is increasing. To improve the accuracy and efficiency of recognizing young red pears, this study proposes an improved model based on the lightweight YOLOv9s, termed YOLOv9s-Pear. By [...] Read more.
With the advancement of computer vision technology, the demand for fruit recognition in agricultural automation is increasing. To improve the accuracy and efficiency of recognizing young red pears, this study proposes an improved model based on the lightweight YOLOv9s, termed YOLOv9s-Pear. By constructing a feature-rich and diverse image dataset of young red pears and introducing spatial-channel decoupled downsampling (SCDown), C2FUIBELAN, and the YOLOv10 detection head (v10detect) modules, the YOLOv9s model was enhanced to achieve efficient recognition of small targets in resource-constrained agricultural environments. Images of young red pears were captured at different times and locations and underwent preprocessing to establish a high-quality dataset. For model improvements, this study integrated the general inverted bottleneck blocks from C2f and MobileNetV4 with the RepNCSPELAN4 module from the YOLOv9s model to form the new C2FUIBELAN module, enhancing the model’s accuracy and training speed for small-scale object detection. Additionally, the SCDown and v10detect modules replaced the original AConv and detection head structures of the YOLOv9s model, further improving performance. The experimental results demonstrated that the YOLOv9s-Pear model achieved high detection accuracy in recognizing young red pears, while reducing computational costs and parameters. The detection accuracy, recall, mean precision, and extended mean precision were 0.971, 0.970, 0.991, and 0.848, respectively. These results confirm the efficiency of the SCDown, C2FUIBELAN, and v10detect modules in young red pear recognition tasks. The findings of this study not only provide a fast and accurate technique for recognizing young red pears but also offer a reference for detecting young fruits of other fruit trees, significantly contributing to the advancement of agricultural automation technology. Full article
Show Figures

Figure 1

23 pages, 14939 KiB  
Article
Dead Fish Detection Model Based on DD-IYOLOv8
by Jianhua Zheng, Yusha Fu, Ruolin Zhao, Junde Lu and Shuangyin Liu
Fishes 2024, 9(9), 356; https://doi.org/10.3390/fishes9090356 (registering DOI) - 12 Sep 2024
Abstract
In aquaculture, the presence of dead fish on the water surface can serve as a bioindicator of health issues or environmental stressors. To enhance the precision of detecting dead fish floating on the water’s surface, this paper proposes a detection approach that integrates [...] Read more.
In aquaculture, the presence of dead fish on the water surface can serve as a bioindicator of health issues or environmental stressors. To enhance the precision of detecting dead fish floating on the water’s surface, this paper proposes a detection approach that integrates data-driven insights with advanced modeling techniques. Firstly, to reduce the influence of aquatic disturbances and branches during the identification process, prior information, such as branches and ripples, is annotated in the dataset to guide the model to better learn the scale and shape characteristics of dead fish, reduce the interference of branch ripples on detection, and thus improve the accuracy of target identification. Secondly, leveraging the foundational YOLOv8 architecture, a DD-IYOLOv8 (Data-Driven Improved YOLOv8) dead fish detection model is designed. Considering the significant changes in the scale of dead fish at different distances, DySnakeConv (Dynamic Snake Convolution) is introduced into the neck network detection head to adaptively adjust the receptive field, thereby improving the network’s capability to capture features. Additionally, a layer for detecting minor objects has been added, and the detection head of YOLOv8 has been modified to 4, allowing the network to better focus on small targets and occluded dead fish, which improves detection performance. Furthermore, the model incorporates a HAM (Hybrid Attention Mechanism) in the later stages of the backbone network to refine global feature extraction, sharpening the model’s focus on dead fish targets and further enhancing detection accuracy. The experimental results showed that the accuracy of DD-IYOLOv8 in detecting dead fish reached 92.8%, the recall rate reached 89.4%, the AP reached 91.7%, and the F1 value reached 91.0%. This study can achieve precise identification of dead fish, which will help promote the research of automatic pond patrol machine ships. Full article
Show Figures

Figure 1

16 pages, 4676 KiB  
Article
Lightweight Substation Equipment Defect Detection Algorithm for Small Targets
by Jianqiang Wang, Yiwei Sun, Ying Lin and Ke Zhang
Sensors 2024, 24(18), 5914; https://doi.org/10.3390/s24185914 (registering DOI) - 12 Sep 2024
Abstract
Substation equipment defect detection has always played an important role in equipment operation and maintenance. However, the task scenarios of substation equipment defect detection are complex and different. Recent studies have revealed issues such as a significant missed detection rate for small-sized targets [...] Read more.
Substation equipment defect detection has always played an important role in equipment operation and maintenance. However, the task scenarios of substation equipment defect detection are complex and different. Recent studies have revealed issues such as a significant missed detection rate for small-sized targets and diminished detection precision. At the same time, the current mainstream detection algorithms are highly complex, which is not conducive to deployment on resource-constrained devices. In view of the above problems, a small target and lightweight substation main scene equipment defect detection algorithm is proposed: Efficient Attentional Lightweight-YOLO (EAL-YOLO), which detection accuracy exceeds the current mainstream model, and the number of parameters and floating point operations (FLOPs) are also advantageous. Firstly, the EfficientFormerV2 is used to optimize the model backbone, and the Large Separable Kernel Attention (LSKA) mechanism has been incorporated into the Spatial Pyramid Pooling Fast (SPPF) to enhance the model’s feature extraction capabilities; secondly, a small target neck network Attentional scale Sequence Fusion P2-Neck (ASF2-Neck) is proposed to enhance the model’s ability to detect small target defects; finally, in order to facilitate deployment on resource-constrained devices, a lightweight shared convolution detection head module Lightweight Shared Convolutional Head (LSCHead) is proposed. Experiments show that compared with YOLOv8n, EAL-YOLO has improved its accuracy by 2.93 percentage points, and the mAP50 of 12 types of typical equipment defects has reached 92.26%. Concurrently, the quantity of FLOPs and parameters has diminished by 46.5% and 61.17% respectively, in comparison with YOLOv8s, meeting the needs of substation defect detection. Full article
Show Figures

Figure 1

14 pages, 3163 KiB  
Article
QYOLO: Contextual Query-Assisted Object Detection in High-Resolution Images
by Mingyang Gao, Wenrui Wang, Jia Mao, Jun Xiong, Zhenming Wang and Bo Wu
Information 2024, 15(9), 563; https://doi.org/10.3390/info15090563 - 12 Sep 2024
Abstract
High-resolution imagery captured by drones can detect critical components on high-voltage transmission towers, providing inspection personnel with essential maintenance insights and improving the efficiency of power line inspections. The high-resolution imagery is particularly effective in enhancing the detection of fine details such as [...] Read more.
High-resolution imagery captured by drones can detect critical components on high-voltage transmission towers, providing inspection personnel with essential maintenance insights and improving the efficiency of power line inspections. The high-resolution imagery is particularly effective in enhancing the detection of fine details such as screws. The QYOLO algorithm, an enhancement of YOLOv8, incorporates context queries into the feature pyramid, effectively capturing long-range dependencies and improving the network’s ability to detect objects. To address the increased network depth and computational load introduced by query extraction, Ghost Separable Convolution (GSConv) is employed, reducing the computational expense by half and further improving the detection performance for small objects such as screws. The experimental validation using the Transmission Line Accessories Dataset (TLAD) developed for this project demonstrates that the proposed improvements increase the average precision (AP) for small objects by 5.5% and the F1-score by 3.5%. The method also enhances detection performance for overall targets, confirming its efficacy in practical applications. Full article
Show Figures

Figure 1

19 pages, 20386 KiB  
Article
YOD-SLAM: An Indoor Dynamic VSLAM Algorithm Based on the YOLOv8 Model and Depth Information
by Yiming Li, Yize Wang, Liuwei Lu and Qi An
Electronics 2024, 13(18), 3633; https://doi.org/10.3390/electronics13183633 - 12 Sep 2024
Abstract
Aiming at the problems of low positioning accuracy and poor mapping effect of the visual SLAM system caused by the poor quality of the dynamic object mask in an indoor dynamic environment, an indoor dynamic VSLAM algorithm based on the YOLOv8 model and [...] Read more.
Aiming at the problems of low positioning accuracy and poor mapping effect of the visual SLAM system caused by the poor quality of the dynamic object mask in an indoor dynamic environment, an indoor dynamic VSLAM algorithm based on the YOLOv8 model and depth information (YOD-SLAM) is proposed based on the ORB-SLAM3 system. Firstly, the YOLOv8 model obtains the original mask of a priori dynamic objects, and the depth information is used to modify the mask. Secondly, the mask’s depth information and center point are used to a priori determine if the dynamic object has missed detection and if the mask needs to be redrawn. Then, the mask edge distance and depth information are used to judge the movement state of non-prior dynamic objects. Finally, all dynamic object information is removed, and the remaining static objects are used for posing estimation and dense point cloud mapping. The accuracy of camera positioning and the construction effect of dense point cloud maps are verified using the TUM RGB-D dataset and real environment data. The results show that YOD-SLAM has a higher positioning accuracy and dense point cloud mapping effect in dynamic scenes than other advanced SLAM systems such as DS-SLAM and DynaSLAM. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

18 pages, 6748 KiB  
Article
FD-Net: A Single-Stage Fire Detection Framework for Remote Sensing in Complex Environments
by Jianye Yuan, Haofei Wang, Minghao Li, Xiaohan Wang, Weiwei Song, Song Li and Wei Gong
Remote Sens. 2024, 16(18), 3382; https://doi.org/10.3390/rs16183382 - 11 Sep 2024
Viewed by 153
Abstract
Fire detection is crucial due to the exorbitant annual toll on both human lives and the economy resulting from fire-related incidents. To enhance forest fire detection in complex environments, we propose a new algorithm called FD-Net for various environments. Firstly, to improve detection [...] Read more.
Fire detection is crucial due to the exorbitant annual toll on both human lives and the economy resulting from fire-related incidents. To enhance forest fire detection in complex environments, we propose a new algorithm called FD-Net for various environments. Firstly, to improve detection performance, we introduce a Fire Attention (FA) mechanism that utilizes the position information from feature maps. Secondly, to prevent geometric distortion during image cropping, we propose a Three-Scale Pooling (TSP) module. Lastly, we fine-tune the YOLOv5 network and incorporate a new Fire Fusion (FF) module to enhance the network’s precision in identifying fire targets. Through qualitative and quantitative comparisons, we found that FD-Net outperforms current state-of-the-art algorithms in performance on both fire and fire-and-smoke datasets. This further demonstrates FD-Net’s effectiveness for application in fire detection. Full article
Show Figures

Figure 1

17 pages, 6083 KiB  
Article
GFI-YOLOv8: Sika Deer Posture Recognition Target Detection Method Based on YOLOv8
by He Gong, Jingyi Liu, Zhipeng Li, Hang Zhu, Lan Luo, Haoxu Li, Tianli Hu, Ying Guo and Ye Mu
Animals 2024, 14(18), 2640; https://doi.org/10.3390/ani14182640 - 11 Sep 2024
Viewed by 230
Abstract
As the sika deer breeding industry flourishes on a large scale, accurately assessing the health of these animals is of paramount importance. Implementing posture recognition through target detection serves as a vital method for monitoring the well-being of sika deer. This approach allows [...] Read more.
As the sika deer breeding industry flourishes on a large scale, accurately assessing the health of these animals is of paramount importance. Implementing posture recognition through target detection serves as a vital method for monitoring the well-being of sika deer. This approach allows for a more nuanced understanding of their physical condition, ensuring the industry can maintain high standards of animal welfare and productivity. In order to achieve remote monitoring of sika deer without interfering with the natural behavior of the animals, and to enhance animal welfare, this paper proposes a sika deer individual posture recognition detection algorithm GFI-YOLOv8 based on YOLOv8. Firstly, this paper proposes to add the iAFF iterative attention feature fusion module to the C2f of the backbone network module, replace the original SPPF module with AIFI module, and use the attention mechanism to adjust the feature channel adaptively. This aims to enhance granularity, improve the model’s recognition, and enhance understanding of sika deer behavior in complex scenes. Secondly, a novel convolutional neural network module is introduced to improve the efficiency and accuracy of feature extraction, while preserving the model’s depth and diversity. In addition, a new attention mechanism module is proposed to expand the receptive field and simplify the model. Furthermore, a new pyramid network and an optimized detection head module are presented to improve the recognition and interpretation of sika deer postures in intricate environments. The experimental results demonstrate that the model achieves 91.6% accuracy in recognizing the posture of sika deer, with a 6% improvement in accuracy and a 4.6% increase in mAP50 compared to YOLOv8n. Compared to other models in the YOLO series, such as YOLOv5n, YOLOv7-tiny, YOLOv8n, YOLOv8s, YOLOv9, and YOLOv10, this model exhibits higher accuracy, and improved mAP50 and mAP50-95 values. The overall performance is commendable, meeting the requirements for accurate and rapid identification of the posture of sika deer. This model proves beneficial for the precise and real-time monitoring of sika deer posture in complex breeding environments and under all-weather conditions. Full article
(This article belongs to the Section Animal System and Management)
Show Figures

Figure 1

15 pages, 11549 KiB  
Article
Automatic Disease Detection from Strawberry Leaf Based on Improved YOLOv8
by Yuelong He, Yunfeng Peng, Chuyong Wei, Yuda Zheng, Changcai Yang and Tengyue Zou
Plants 2024, 13(18), 2556; https://doi.org/10.3390/plants13182556 - 11 Sep 2024
Viewed by 240
Abstract
Strawberries are susceptible to various diseases during their growth, and leaves may show signs of diseases as a response. Given that these diseases generate yield loss and compromise the quality of strawberries, timely detection is imperative. To automatically identify diseases in strawberry leaves, [...] Read more.
Strawberries are susceptible to various diseases during their growth, and leaves may show signs of diseases as a response. Given that these diseases generate yield loss and compromise the quality of strawberries, timely detection is imperative. To automatically identify diseases in strawberry leaves, a KTD-YOLOv8 model is introduced to enhance both accuracy and speed. The KernelWarehouse convolution is employed to replace the traditional component in the backbone of the YOLOv8 to reduce the computational complexity. In addition, the Triplet Attention mechanism is added to fully extract and fuse multi-scale features. Furthermore, a parameter-sharing diverse branch block (DBB) sharing head is constructed to improve the model’s target processing ability at different spatial scales and increase its accuracy without adding too much calculation. The experimental results show that, compared with the original YOLOv8, the proposed KTD-YOLOv8 increases the average accuracy by 2.8% and reduces the floating-point calculation by 38.5%. It provides a new option to guide the intelligent plant monitoring system and precision pesticide spraying system during the growth of strawberry plants. Full article
(This article belongs to the Section Plant Modeling)
Show Figures

Figure 1

12 pages, 3551 KiB  
Article
Deep Learning Based Characterization of Cold-Water Coral Habitat at Central Cantabrian Natura 2000 Sites Using YOLOv8
by Alberto Gayá-Vilar, Alberto Abad-Uribarren, Augusto Rodríguez-Basalo, Pilar Ríos, Javier Cristobo and Elena Prado
J. Mar. Sci. Eng. 2024, 12(9), 1617; https://doi.org/10.3390/jmse12091617 - 11 Sep 2024
Viewed by 148
Abstract
Cold-water coral (CWC) reefs, such as those formed by Desmophyllum pertusum and Madrepora oculata, are vital yet vulnerable marine ecosystems (VMEs). The need for accurate and efficient monitoring of these habitats has driven the exploration of innovative approaches. This study presents a [...] Read more.
Cold-water coral (CWC) reefs, such as those formed by Desmophyllum pertusum and Madrepora oculata, are vital yet vulnerable marine ecosystems (VMEs). The need for accurate and efficient monitoring of these habitats has driven the exploration of innovative approaches. This study presents a novel application of the YOLOv8l-seg deep learning model for the automated detection and segmentation of these key CWC species in underwater imagery. The model was trained and validated on images collected at two Natura 2000 sites in the Cantabrian Sea: the Avilés Canyon System (ACS) and El Cachucho Seamount (CSM). Results demonstrate the model’s high accuracy in identifying and delineating individual coral colonies, enabling the assessment of coral cover and spatial distribution. The study revealed significant variability in coral cover between and within the study areas, highlighting the patchy nature of CWC habitats. Three distinct coral community groups were identified based on percentage coverage composition and abundance, with the highest coral cover group being located exclusively in the La Gaviera canyon head within the ACS. This research underscores the potential of deep learning models for efficient and accurate monitoring of VMEs, facilitating the acquisition of high-resolution data essential for understanding CWC distribution, abundance, and community structure, and ultimately contributing to the development of effective conservation strategies. Full article
(This article belongs to the Special Issue Application of Deep Learning in Underwater Image Processing)
Show Figures

Figure 1

18 pages, 15545 KiB  
Article
YOLO-BGS Optimizes Textile Production Processes: Enhancing YOLOv8n with Bi-Directional Feature Pyramid Network and Global and Shuffle Attention Mechanisms for Efficient Fabric Defect Detection
by Gege Lu, Tian Xiong and Gaihong Wu
Sustainability 2024, 16(18), 7922; https://doi.org/10.3390/su16187922 - 11 Sep 2024
Viewed by 238
Abstract
Timely detection of fabric defects is crucial for improving fabric quality and reducing production losses for companies. Traditional methods for detecting fabric defects face several challenges, including low detection efficiency, poor accuracy, and limited types of detectable defects. To address these issues, this [...] Read more.
Timely detection of fabric defects is crucial for improving fabric quality and reducing production losses for companies. Traditional methods for detecting fabric defects face several challenges, including low detection efficiency, poor accuracy, and limited types of detectable defects. To address these issues, this paper chose the YOLOv8n model for continuous iteration enhancement in order to improve its detection performance. First, multiscale feature fusion was realized by the Bi-directional Feature Pyramid Network (BiFPN). Second, the Shuffle Attention Mechanism (SA) is introduced to optimize feature classification. Finally, the Global Attention Mechanism (GAM) was used to improve global detection accuracy. Empirical findings demonstrated the improved model’s efficacy, attaining a test set mean average precision (mAP) value of 96.6%, which is an improvement of 3.6% compared to the original YOLOv8n. This validates that YOLO-BGS excels in detecting textile defects. It effectively locates these defects, minimizes resource waste, and fosters sustainable production practices. Full article
Show Figures

Figure 1

13 pages, 5820 KiB  
Article
Optic Nerve Sheath Ultrasound Image Segmentation Based on CBC-YOLOv5s
by Yonghua Chu, Jinyang Xu, Chunshuang Wu, Jianping Ye, Jucheng Zhang, Lei Shen, Huaxia Wang and Yudong Yao
Electronics 2024, 13(18), 3595; https://doi.org/10.3390/electronics13183595 - 10 Sep 2024
Viewed by 227
Abstract
The diameter of the optic nerve sheath is an important indicator for assessing the intracranial pressure in critically ill patients. The methods for measuring the optic nerve sheath diameter are generally divided into invasive and non-invasive methods. Compared to the invasive methods, the [...] Read more.
The diameter of the optic nerve sheath is an important indicator for assessing the intracranial pressure in critically ill patients. The methods for measuring the optic nerve sheath diameter are generally divided into invasive and non-invasive methods. Compared to the invasive methods, the non-invasive methods are safer and have thus gained popularity. Among the non-invasive methods, using deep learning to process the ultrasound images of the eyes of critically ill patients and promptly output the diameter of the optic nerve sheath offers significant advantages. This paper proposes a CBC-YOLOv5s optic nerve sheath ultrasound image segmentation method that integrates both local and global features. First, it introduces the CBC-Backbone feature extraction network, which consists of dual-layer C3 Swin-Transformer (C3STR) and dual-layer Bottleneck Transformer (BoT3) modules. The C3STR backbone’s multi-layer convolution and residual connections focus on the local features of the optic nerve sheath, while the Window Transformer Attention (WTA) mechanism in the C3STR module and the Multi-Head Self-Attention (MHSA) in the BoT3 module enhance the model’s understanding of the global features of the optic nerve sheath. The extracted local and global features are fully integrated in the Spatial Pyramid Pooling Fusion (SPPF) module. Additionally, the CBC-Neck feature pyramid is proposed, which includes a single-layer C3STR module and three-layer CReToNeXt (CRTN) module. During upsampling feature fusion, the C3STR module is used to enhance the local and global awareness of the fused features. During downsampling feature fusion, the CRTN module’s multi-level residual design helps the network to better capture the global features of the optic nerve sheath within the fused features. The introduction of these modules achieves the thorough integration of the local and global features, enabling the model to efficiently and accurately identify the optic nerve sheath boundaries, even when the ocular ultrasound images are blurry or the boundaries are unclear. The Z2HOSPITAL-5000 dataset collected from Zhejiang University Second Hospital was used for the experiments. Compared to the widely used YOLOv5s and U-Net algorithms, the proposed method shows improved performance on the blurry test set. Specifically, the proposed method achieves precision, recall, and Intersection over Union (IoU) values that are 4.1%, 2.1%, and 4.5% higher than those of YOLOv5s. When compared to U-Net, the precision, recall, and IoU are improved by 9.2%, 21%, and 19.7%, respectively. Full article
(This article belongs to the Special Issue Deep Learning-Based Object Detection/Classification)
Show Figures

Figure 1

19 pages, 10176 KiB  
Article
A Lightweight Real-Time Recognition Algorithm for Tomato Leaf Disease Based on Improved YOLOv8
by Wenbo Liu, Chenhao Bai, Wei Tang, Yu Xia and Jie Kang
Agronomy 2024, 14(9), 2069; https://doi.org/10.3390/agronomy14092069 - 10 Sep 2024
Viewed by 175
Abstract
To address the real-time detection challenge of deploying deep learning-based tomato leaf disease detection algorithms on embedded devices, an improved tomato leaf disease detection algorithm based on YOLOv8n is proposed in this paper. It is able to achieve the efficient, real-time detection of [...] Read more.
To address the real-time detection challenge of deploying deep learning-based tomato leaf disease detection algorithms on embedded devices, an improved tomato leaf disease detection algorithm based on YOLOv8n is proposed in this paper. It is able to achieve the efficient, real-time detection of tomato leaf diseases while maintaining model’s lightweight requirements. The algorithm incorporated the LMSM (lightweight multi-scale module) and ALSA (Attention Lightweight Subsampling Module) to improve the ability to extract lightweight and multi-scale semantic information for the specific characteristics of tomato leaf disease, which include irregular spot size and lush tomato leaves. The head network was redesigned utilizing partial and group convolution along with a parameter-sharing method. Scalable auxiliary bounding box and loss function optimization strategies were introduced to further enhance performance. After undergoing the pruning technique, computation decreased by 61.7%, the model size decreased by 55.6%, and the FPS increased by 44.8%, all while a high level of accuracy was maintained. A detection speed of 19.70FPS on the Jetson Nano was obtained after undergoing TensorRT quantization, showing a 64.85% improvement compared to the initial detection speed. This method met the high real-time performance and small model size requirements for embedded tomato leaf disease detection systems, indirectly reducing the energy consumption of online detection. It provided an effective solution for the online detection of tomato leaf disease. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

15 pages, 4200 KiB  
Article
Research on Rail Surface Defect Detection Based on Improved CenterNet
by Yizhou Mao, Shubin Zheng, Liming Li, Renjie Shi and Xiaoxue An
Electronics 2024, 13(17), 3580; https://doi.org/10.3390/electronics13173580 - 9 Sep 2024
Viewed by 415
Abstract
Rail surface defect detection is vital for railway safety. Traditional methods falter with varying defect sizes and complex backgrounds, while two-stage deep learning models, though accurate, lack real-time capabilities. To overcome these challenges, we propose an enhanced one-stage detection model based on CenterNet. [...] Read more.
Rail surface defect detection is vital for railway safety. Traditional methods falter with varying defect sizes and complex backgrounds, while two-stage deep learning models, though accurate, lack real-time capabilities. To overcome these challenges, we propose an enhanced one-stage detection model based on CenterNet. We replace ResNet with ResNeXt and implement a multi-branch structure for better low-level feature extraction. Additionally, we integrate SKNet attention mechanism with the C2f structure from YOLOv8, improving the model’s focus on critical image regions and enhancing the detection of minor defects. We also introduce an elliptical Gaussian kernel for size regression loss, better representing the aspect ratio of rail defects. This approach enhances detection accuracy and speeds up training. Our model achieves a mean accuracy (mAP) of 0.952 on the rail defects dataset, outperforming other models with a 6.6% improvement over the original and a 35.5% increase in training speed. These results demonstrate the efficiency and reliability of our method for rail defect detection. Full article
Show Figures

Figure 1

26 pages, 4456 KiB  
Article
Generating 3D Models for UAV-Based Detection of Riparian PET Plastic Bottle Waste: Integrating Local Social Media and InstantMesh
by Shijun Pan, Keisuke Yoshida, Daichi Shimoe, Takashi Kojima and Satoshi Nishiyama
Viewed by 362
Abstract
In recent years, waste pollution has become a severe threat to riparian environments worldwide. Along with the advancement of deep learning (DL) algorithms (i.e., object detection models), related techniques have become useful for practical applications. This work attempts to develop a data generation [...] Read more.
In recent years, waste pollution has become a severe threat to riparian environments worldwide. Along with the advancement of deep learning (DL) algorithms (i.e., object detection models), related techniques have become useful for practical applications. This work attempts to develop a data generation approach to generate datasets for small target recognition, especially for recognition in remote sensing images. A relevant point is that similarity between data used for model training and data used for testing is crucially important for object detection model performance. Therefore, obtaining training data with high similarity to the monitored objects is a key objective of this study. Currently, Artificial Intelligence Generated Content (AIGC), such as single target objects generated by Luma AI, is a promising data source for DL-based object detection models. However, most of the training data supporting the generated results are not from Japan. Consequently, the generated data are less similar to monitored objects in Japan, having, for example, different label colors, shapes, and designs. For this study, the authors developed a data generation approach by combining social media (Clean-Up Okayama) and single-image-based 3D model generation algorithms (e.g., InstantMesh) to provide a reliable reference for future generations of localized data. The trained YOLOv8 model in this research, obtained from the S2PS (Similar to Practical Situation) AIGC dataset, produced encouraging results (high F1 scores, approximately 0.9) in scenario-controlled UAV-based riparian PET bottle waste identification tasks. The results of this study show the potential of AIGC to supplement or replace real-world data collection and reduce the on-site work load. Full article
15 pages, 6691 KiB  
Article
Engineering Vehicle Detection Based on Improved YOLOv6
by Huixuan Ling, Tianju Zhao, Yangqianhui Zhang and Meng Lei
Appl. Sci. 2024, 14(17), 8054; https://doi.org/10.3390/app14178054 - 9 Sep 2024
Viewed by 277
Abstract
Engineering vehicles play a vital role in supporting construction projects. However, due to their substantial size, heavy tonnage, and significant blind spots while in motion, they present a potential threat to road maintenance, pedestrian safety, and the well-being of other vehicles. Hence, monitoring [...] Read more.
Engineering vehicles play a vital role in supporting construction projects. However, due to their substantial size, heavy tonnage, and significant blind spots while in motion, they present a potential threat to road maintenance, pedestrian safety, and the well-being of other vehicles. Hence, monitoring engineering vehicles holds considerable importance. This paper introduces an engineering vehicle detection model based on improved YOLOv6. First, a Swin Transformer is employed for feature extraction, capturing comprehensive image features to improve the detection capability of incomplete objects. Subsequently, the SimMIM self-supervised training paradigm is implemented to address challenges related to insufficient data and high labeling costs. Experimental results demonstrate the model’s superior performance, with a mAP50:95 value of 88.5% and mAP50 value of 95.9% on the dataset of four types of engineering vehicles, surpassing existing mainstream models and proving its effectiveness in engineering vehicle detection. Full article
Show Figures

Figure 1

Back to TopTop