Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (3,894)

Search Parameters:
Keywords = ensemble methods

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
24 pages, 3706 KiB  
Article
A Sequence-Based Hybrid Ensemble Approach for Estimating Trail Pavement Roughness Using Smartphone and Bicycle Data
by Yazan Ibrahim Alatoom, Zia U. Zihan, Inya Nlenanya, Abdallah B. Al-Hamdan and Omar Smadi
Infrastructures 2024, 9(10), 179; https://doi.org/10.3390/infrastructures9100179 - 8 Oct 2024
Abstract
Trail pavement roughness significantly impacts user experience and safety. Measuring roughness over large areas using traditional equipment is challenging and expensive. The utilization of smartphones and bicycles offers a more feasible approach to measuring trail roughness, but the current methods to capture data [...] Read more.
Trail pavement roughness significantly impacts user experience and safety. Measuring roughness over large areas using traditional equipment is challenging and expensive. The utilization of smartphones and bicycles offers a more feasible approach to measuring trail roughness, but the current methods to capture data using these have accuracy limitations. While machine learning has the potential to improve accuracy, there have been few applications of real-time roughness evaluation. This study proposes a hybrid ensemble machine learning model that combines sequence-based modeling with support vector regression (SVR) to estimate trail roughness using smartphone sensor data mounted on bicycles. The hybrid model outperformed traditional methods like double integration and whole-body vibration in roughness estimation. For the 0.031 mi (50 m) segments, it reduced RMSE by 54–74% for asphalt concrete (AC) trails and 50–59% for Portland cement concrete (PCC) trails. For the 0.31 mi (499 m) segments, RMSE reductions of 37–60% and 49–56% for AC and PCC trails were achieved, respectively. Additionally, the hybrid model outperformed the base random forest model by 17%, highlighting the effectiveness of combining ensemble learning with sequence modeling and SVR. These results demonstrate that the hybrid model provides a cost-effective, scalable, and highly accurate alternative for large-scale trail roughness monitoring and assessment. Full article
(This article belongs to the Special Issue Pavement Design and Pavement Management)
25 pages, 5210 KiB  
Article
Application of SHAP and Multi-Agent Approach for Short-Term Forecast of Power Consumption of Gas Industry Enterprises
by Alina I. Stepanova, Alexandra I. Khalyasmaa, Pavel V. Matrenin and Stanislav A. Eroshenko
Algorithms 2024, 17(10), 447; https://doi.org/10.3390/a17100447 - 8 Oct 2024
Abstract
Currently, machine learning methods are widely applied in the power industry to solve various tasks, including short-term power consumption forecasting. However, the lack of interpretability of machine learning methods can lead to their incorrect use, potentially resulting in electrical system instability or equipment [...] Read more.
Currently, machine learning methods are widely applied in the power industry to solve various tasks, including short-term power consumption forecasting. However, the lack of interpretability of machine learning methods can lead to their incorrect use, potentially resulting in electrical system instability or equipment failures. This article addresses the task of short-term power consumption forecasting, one of the tasks of enhancing the energy efficiency of gas industry enterprises. In order to reduce the risks of making incorrect decisions based on the results of short-term power consumption forecasts made by machine learning methods, the SHapley Additive exPlanations method was proposed. Additionally, the application of a multi-agent approach for the decomposition of production processes using self-generation agents, energy storage agents, and consumption agents was demonstrated. It can enable the safe operation of critical infrastructure, for instance, adjusting the operation modes of self-generation units and energy-storage systems, optimizing the power consumption schedule, and reducing electricity and power costs. A comparative analysis of various algorithms for constructing decision tree ensembles was conducted to forecast power consumption by gas industry enterprises with different numbers of categorical features. The experiments demonstrated that using the developed method and production process factors reduced the MAE from 105.00 kWh (MAPE of 16.81%), obtained through expert forecasting, to 15.52 kWh (3.44%). Examples were provided of how the use of SHapley Additive exPlanation can increase the safety of the electrical system management of gas industry enterprises by improving experts’ confidence in the results of the information system. Full article
Show Figures

Figure 1

21 pages, 9608 KiB  
Article
Ensemble Machine-Learning-Based Framework for Estimating Surface Soil Moisture Using Sentinel-1/2 Data: A Case Study of an Arid Oasis in China
by Junhao Liu, Zhe Hao, Jianli Ding, Yukun Zhang, Zhiguo Miao, Yu Zheng, Alimira Alimu, Huiling Cheng and Xiang Li
Land 2024, 13(10), 1635; https://doi.org/10.3390/land13101635 - 8 Oct 2024
Abstract
Soil moisture (SM) is a critical parameter in Earth’s water cycle, significantly impacting hydrological, agricultural, and meteorological research fields. The challenge of estimating surface soil moisture from synthetic aperture radar (SAR) data is compounded by the influence of vegetation coverage. This study focuses [...] Read more.
Soil moisture (SM) is a critical parameter in Earth’s water cycle, significantly impacting hydrological, agricultural, and meteorological research fields. The challenge of estimating surface soil moisture from synthetic aperture radar (SAR) data is compounded by the influence of vegetation coverage. This study focuses on the Weigan River and Kuche River Delta Oasis in Xinjiang, employing high-resolution Sentinel-1 and Sentinel-2 images in conjunction with a modified Water Cloud Model (WCM) and the grayscale co-occurrence matrix (GLCM) for feature parameter extraction. A soil moisture inversion method based on stacked ensemble learning is proposed, which integrates random forest, CatBoost, and LightGBM. The findings underscore the feasibility of using multi-source remote sensing data for oasis moisture inversion in arid regions. However, soil moisture content estimates tend to be overestimated above 10% and underestimated below 5%. The CatBoost model achieved the highest accuracy (R2 = 0.827, RMSE = 0.014 g/g) using the top 16 feature parameter groups. Additionally, the R2 values for Stacking1 and Stacking2 models saw increases of 0.008 and 0.016, respectively. Thus, integrating multi-source remote sensing data with Stacking models offers valuable support and reference for large-scale estimation of surface soil moisture content in arid oasis areas. Full article
Show Figures

Figure 1

27 pages, 2585 KiB  
Article
Technology-Driven Financial Risk Management: Exploring the Benefits of Machine Learning for Non-Profit Organizations
by Hao Huang
Systems 2024, 12(10), 416; https://doi.org/10.3390/systems12100416 - 8 Oct 2024
Viewed by 96
Abstract
This study explores how machine learning can optimize financial risk management for non-profit organizations by evaluating various algorithms aimed at mitigating loan default risks. The findings indicate that ensemble learning models, such as random forest and LightGBM, significantly improve prediction accuracy, thereby enabling [...] Read more.
This study explores how machine learning can optimize financial risk management for non-profit organizations by evaluating various algorithms aimed at mitigating loan default risks. The findings indicate that ensemble learning models, such as random forest and LightGBM, significantly improve prediction accuracy, thereby enabling non-profits to better manage financial risk. In the context of the 2008 subprime mortgage crisis, which underscored the volatility of financial markets, this research assesses a range of risks—credit, operational, liquidity, and market risks—while exploring both traditional machine learning and advanced ensemble techniques, with a particular focus on stacking fusion to enhance model performance. Emphasizing the importance of privacy and adaptive methods, this study advocates for interdisciplinary approaches to overcome limitations such as stress testing, data analysis rule formulation, and regulatory collaboration. The research underscores machine learning’s crucial role in financial risk control and calls on regulatory authorities to reassess existing frameworks to accommodate evolving risks. Additionally, it highlights the need for accurate data type identification and the potential for machine learning to strengthen financial risk management amid uncertainty, promoting interdisciplinary efforts that address broader issues like environmental sustainability and economic development. Full article
Show Figures

Figure 1

14 pages, 526 KiB  
Article
Assessment of Ensemble-Based Machine Learning Algorithms for Exoplanet Identification
by Thiago S. F. Luz, Rodrigo A. S. Braga and Enio R. Ribeiro
Electronics 2024, 13(19), 3950; https://doi.org/10.3390/electronics13193950 - 7 Oct 2024
Viewed by 306
Abstract
This paper presents a comprehensive assessment procedure for evaluating Ensemble-based Machine Learning algorithms in the context of exoplanet classification. Each of the algorithm hyperparameter values were tuned. Deployments were carried out using the cross-validation method. Performance metrics, including accuracy, sensitivity, specificity, precision, and [...] Read more.
This paper presents a comprehensive assessment procedure for evaluating Ensemble-based Machine Learning algorithms in the context of exoplanet classification. Each of the algorithm hyperparameter values were tuned. Deployments were carried out using the cross-validation method. Performance metrics, including accuracy, sensitivity, specificity, precision, and F1 score, were evaluated using confusion matrices generated from each implementation. Machine Learning (ML) algorithms were trained and used to identify exoplanet data. Most of the current research deals with traditional ML algorithms for this purpose. The Ensemble algorithm is another type of ML technique that combines the prediction performance of two or more algorithms to obtain an improved final prediction. Few studies have applied Ensemble algorithms to predict exoplanets. To the best of our knowledge, no paper that has exclusively assessed Ensemble algorithms exists, highlighting a significant gap in the literature about the potential of Ensemble methods. Five Ensemble algorithms were evaluated in this paper: Adaboost, Random Forest, Stacking, Random Subspace Method, and Extremely Randomized Trees. They achieved an average performance of more than 80% in all metrics. The results underscore the substantial benefits of fine tuning hyperparameters to enhance predictive performance. The Stacking algorithm achieved a higher performance than the other algorithms. This aspect is discussed in this paper. The results of this work show that it is worth increasing the use of Ensemble algorithms to improve exoplanet identification. Full article
(This article belongs to the Section Computer Science & Engineering)
Show Figures

Figure 1

14 pages, 1842 KiB  
Article
Prediction of Flotation Deinking Performance: A Comparative Analysis of Machine Learning Techniques
by Tamara Gavrilović, Vladimir Despotović, Madalina-Ileana Zot and Maja S. Trumić
Appl. Sci. 2024, 14(19), 8990; https://doi.org/10.3390/app14198990 - 6 Oct 2024
Viewed by 369
Abstract
Flotation deinking is one of the most widely used techniques for the separation of ink particles from cellulose fibers during the process of paper recycling. It is a complex process influenced by a variety of factors, and is difficult to represent and usually [...] Read more.
Flotation deinking is one of the most widely used techniques for the separation of ink particles from cellulose fibers during the process of paper recycling. It is a complex process influenced by a variety of factors, and is difficult to represent and usually results in models that are inconvenient to implement and/or interpret. In this paper, a comprehensive study of several machine learning methods for the prediction of flotation deinking performance is carried out, including support vector regression, regression tree ensembles (random forests and boosting) and Gaussian process regression. The prediction relies on the development of a limited dataset that assumes representative data samples obtained under a variety of laboratory conditions, including different reagents, pH values and flotation residence times. The results obtained in this paper confirm that the machine learning methods enable the accurate prediction of flotation deinking performance even when the dataset used for training the model is limited, thus enabling the determination of optimal conditions for the paper recycling process, with only minimal costs and effort. Considering the low complexity of the Gaussian process regression compared to the aforementioned ensemble models, it should be emphasized that the Gaussian process regression gave the best performance in estimating fiber recovery (R2 = 97.77%) and a reasonable performance in estimating the toner recovery (R2 = 86.31%). Full article
Show Figures

Figure 1

21 pages, 1898 KiB  
Article
Machine Learning Valuation in Dual Market Dynamics: A Case Study of the Formal and Informal Real Estate Market in Dar es Salaam
by Frank Nyanda, Henry Muyingo and Mats Wilhelmsson
Buildings 2024, 14(10), 3172; https://doi.org/10.3390/buildings14103172 - 5 Oct 2024
Viewed by 450
Abstract
The housing market in Dar es Salaam, Tanzania, is expanding and with it a need for increased market transparency to guide investors and other stakeholders. The objective of this paper is to evaluate machine learning (ML) methods to appraise real estate in formal [...] Read more.
The housing market in Dar es Salaam, Tanzania, is expanding and with it a need for increased market transparency to guide investors and other stakeholders. The objective of this paper is to evaluate machine learning (ML) methods to appraise real estate in formal and informal housing markets in this nascent market sector. Various advanced ML models are applied with the aim of improving property value estimates in a market with limited access to information. The dataset used included detailed property characteristics and transaction data from both market types. Regression, decision trees, neural networks, and ensemble methods were employed to refine property appraisals across these settings. The findings indicate significant differences between formal and informal market valuations, demonstrating ML’s effectiveness in handling limited data and complex market dynamics. These results emphasise the potential of ML techniques in emerging markets where traditional valuation methods often fail due to the scarcity of transaction data. Full article
(This article belongs to the Special Issue Housing Price Dynamics and the Property Market)
Show Figures

Figure 1

18 pages, 841 KiB  
Article
Exploiting Content Characteristics for Explainable Detection of Fake News
by Sergio Muñoz and Carlos Á. Iglesias
Big Data Cogn. Comput. 2024, 8(10), 129; https://doi.org/10.3390/bdcc8100129 - 4 Oct 2024
Viewed by 496
Abstract
The proliferation of fake news threatens the integrity of information ecosystems, creating a pressing need for effective and interpretable detection mechanisms. Recent advances in machine learning, particularly with transformer-based models, offer promising solutions due to their superior ability to analyze complex language patterns. [...] Read more.
The proliferation of fake news threatens the integrity of information ecosystems, creating a pressing need for effective and interpretable detection mechanisms. Recent advances in machine learning, particularly with transformer-based models, offer promising solutions due to their superior ability to analyze complex language patterns. However, the practical implementation of these solutions often presents challenges due to their high computational costs and limited interpretability. In this work, we explore using content-based features to enhance the explainability and effectiveness of fake news detection. We propose a comprehensive feature framework encompassing characteristics related to linguistic, affective, cognitive, social, and contextual processes. This framework is evaluated across several public English datasets to identify key differences between fake and legitimate news. We assess the detection performance of these features using various traditional classifiers, including single and ensemble methods and analyze how feature reduction affects classifier performance. Our results show that, while traditional classifiers may not fully match transformer-based models, they achieve competitive results with significantly lower computational requirements. We also provide an interpretability analysis highlighting the most influential features in classification decisions. This study demonstrates the potential of interpretable features to build efficient, explainable, and accessible fake news detection systems. Full article
Show Figures

Figure 1

37 pages, 11643 KiB  
Article
Co-CrackSegment: A New Collaborative Deep Learning Framework for Pixel-Level Semantic Segmentation of Concrete Cracks
by Nizar Faisal Alkayem, Ali Mayya, Lei Shen, Xin Zhang, Panagiotis G. Asteris, Qiang Wang and Maosen Cao
Mathematics 2024, 12(19), 3105; https://doi.org/10.3390/math12193105 - 4 Oct 2024
Viewed by 395
Abstract
In an era of massive construction, damaged and aging infrastructure are becoming more common. Defects, such as cracking, spalling, etc., are main types of structural damage that widely occur. Hence, ensuring the safe operation of existing infrastructure through health monitoring has emerged as [...] Read more.
In an era of massive construction, damaged and aging infrastructure are becoming more common. Defects, such as cracking, spalling, etc., are main types of structural damage that widely occur. Hence, ensuring the safe operation of existing infrastructure through health monitoring has emerged as an important challenge facing engineers. In recent years, intelligent approaches, such as data-driven machines and deep learning crack detection have gradually dominated over traditional methods. Among them, the semantic segmentation using deep learning models is a process of the characterization of accurate locations and portraits of cracks using pixel-level classification. Most available studies rely on single-model knowledge to perform this task. However, it is well-known that the single model might suffer from low variance and low ability to generalize in case of data alteration. By leveraging the ensemble deep learning philosophy, a novel collaborative semantic segmentation of concrete cracks method called Co-CrackSegment is proposed. Firstly, five models, namely the U-net, SegNet, DeepCrack19, DeepLabV3-ResNet50, and DeepLabV3-ResNet101 are trained to serve as core models for the ensemble model Co-CrackSegment. To build the ensemble model Co-CrackSegment, a new iterative approach based on the best evaluation metrics, namely the Dice score, IoU, pixel accuracy, precision, and recall metrics is developed. Results show that the Co-CrackSegment exhibits a prominent performance compared with core models and weighted average ensemble by means of the considered best statistical metrics. Full article
Show Figures

Figure 1

28 pages, 1573 KiB  
Article
Soil Properties Classification in Sustainable Agriculture Using Genetic Algorithm-Optimized and Deep Neural Networks
by Yadviga Tynchenko, Vadim Tynchenko, Vladislav Kukartsev, Tatyana Panfilova, Oksana Kukartseva, Ksenia Degtyareva, Van Nguyen and Ivan Malashin
Sustainability 2024, 16(19), 8598; https://doi.org/10.3390/su16198598 - 3 Oct 2024
Viewed by 527
Abstract
Optimization of land management and agricultural practices require precise classification of soil properties. This study presents a method to fine-tune deep neural network (DNN) hyperparameters for multiclass classification of soil properties using genetic algorithms (GAs) with knowledge-based generation of hyperparameters. The focus is [...] Read more.
Optimization of land management and agricultural practices require precise classification of soil properties. This study presents a method to fine-tune deep neural network (DNN) hyperparameters for multiclass classification of soil properties using genetic algorithms (GAs) with knowledge-based generation of hyperparameters. The focus is on classifying soil attributes, including nutrient availability (0.78 ± 0.11), nutrient retention capacity (0.86 ± 0.05), rooting conditions (0.85 ± 0.07), oxygen availability to roots (0.84 ± 0.05), excess salts (0.96 ± 0.02), toxicity (0.96 ± 0.01), and soil workability (0.84 ± 0.09), with these accuracies representing the results from classification with variations from cross-validation. A dataset from the USA, which includes land-use distribution, aspect distribution, slope distribution, and climate data for each plot, is utilized. A GA is applied to explore a wide range of hyperparameters, such as the number of layers, neurons per layer, activation functions, optimizers, learning rates, and loss functions. Additionally, ensemble methods such as random forest and gradient boosting machines were employed, demonstrating comparable accuracy to the DNN approach. This research contributes to the advancement of precision agriculture by providing a robust machine learning (ML) framework for accurate soil property classification. By enabling more informed and efficient land management decisions, it promotes sustainable agricultural practices that optimize resource use and enhance soil health for long-term ecological balance. Full article
Show Figures

Figure 1

17 pages, 2356 KiB  
Article
Remaining Useful Life Estimation of Lithium-Ion Batteries Based on Small Sample Models
by Lu Liu, Wei Sun, Chuanxu Yue, Yunhai Zhu and Weihuan Xia
Energies 2024, 17(19), 4932; https://doi.org/10.3390/en17194932 - 2 Oct 2024
Viewed by 423
Abstract
Accurate prediction of the Remaining Useful Life (RUL) of lithium-ion batteries is essential for enhancing energy management and extending the lifespan of batteries across various industries. However, the raw capacity data of these batteries is often noisy and exhibits complex nonlinear degradation patterns, [...] Read more.
Accurate prediction of the Remaining Useful Life (RUL) of lithium-ion batteries is essential for enhancing energy management and extending the lifespan of batteries across various industries. However, the raw capacity data of these batteries is often noisy and exhibits complex nonlinear degradation patterns, especially due to capacity regeneration phenomena during operation, making precise RUL prediction a significant challenge. Although various deep learning-based methods have been proposed, their performance relies heavily on the availability of large datasets, and satisfactory prediction accuracy is often achievable only with extensive training samples. To overcome this limitation, we propose a novel method that integrates sequence decomposition algorithms with an optimized neural network. Specifically, the Complementary Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) algorithm is employed to decompose the raw capacity data, effectively mitigating the noise from capacity regeneration. Subsequently, Particle Swarm Optimization (PSO) is used to fine-tune the hyperparameters of the Bidirectional Gated Recurrent Unit (BiGRU) model. The final BiGRU-based prediction model was extensively tested on eight lithium-ion battery datasets from NASA and CALCE, demonstrating robust generalization capability, even with limited data. The experimental results indicate that the CEEMDAN-PSO-BiGRU model can reliably and accurately predict the RUL and capacity of lithium-ion batteries, providing a promising and reliable method for RUL prediction in practical applications. Full article
(This article belongs to the Section D: Energy Storage and Application)
Show Figures

Figure 1

13 pages, 3660 KiB  
Article
A Novel Surrogate-Assisted Multi-Objective Well Control Parameter Optimization Method Based on Selective Ensembles
by Lian Wang, Rui Deng, Liang Zhang, Jianhua Qu, Hehua Wang, Liehui Zhang, Xing Zhao, Bing Xu, Xindong Lv and Caspar Daniel Adenutsi
Processes 2024, 12(10), 2140; https://doi.org/10.3390/pr12102140 - 1 Oct 2024
Viewed by 487
Abstract
Multi-objective optimization algorithms are crucial for addressing real-world problems, particularly with regard to optimizing well control parameters, which are often computationally expensive due to their reliance on numerical simulations. Surrogate-assisted models help to reduce this computational burden, but their effectiveness depends on the [...] Read more.
Multi-objective optimization algorithms are crucial for addressing real-world problems, particularly with regard to optimizing well control parameters, which are often computationally expensive due to their reliance on numerical simulations. Surrogate-assisted models help to reduce this computational burden, but their effectiveness depends on the quality of the surrogates, which can be affected by candidate dimension and noise. This study proposes a novel surrogate-assisted multi-objective optimization framework (MOO-SESA) that combines selective ensemble support-vector regression with NSGA-II. The framework’s uniqueness lies in its adaptive selection of a diverse subset of surrogates, established prior to iteration, to enhance accuracy, robustness, and computational efficiency. To our knowledge, this is the first instance in which selective ensemble techniques with multi-objective optimization have been applied to reservoir well control problems. Through employing an ensemble strategy for improving the quality of the surrogate model, MOO-SESA demonstrated superior well control scenarios and faster convergence compared to traditional surrogate-assisted models when applied to the SPE10 and Egg reservoir models. Full article
(This article belongs to the Special Issue Advances in Enhancing Unconventional Oil/Gas Recovery, 2nd Edition)
Show Figures

Figure 1

32 pages, 30650 KiB  
Article
A Novel Strategy Coupling Optimised Sampling with Heterogeneous Ensemble Machine-Learning to Predict Landslide Susceptibility
by Yongxing Lu, Honggen Xu, Can Wang, Guanxi Yan, Zhitao Huo, Zuwu Peng, Bo Liu and Chong Xu
Remote Sens. 2024, 16(19), 3663; https://doi.org/10.3390/rs16193663 - 1 Oct 2024
Viewed by 560
Abstract
The accuracy of data-driven landslide susceptibility prediction depends heavily on the quality of non-landslide samples and the selection of machine-learning algorithms. Current methods rely on artificial prior knowledge to obtain negative samples from landslide-free regions or outside the landslide buffer zones randomly and [...] Read more.
The accuracy of data-driven landslide susceptibility prediction depends heavily on the quality of non-landslide samples and the selection of machine-learning algorithms. Current methods rely on artificial prior knowledge to obtain negative samples from landslide-free regions or outside the landslide buffer zones randomly and quickly but often ignore the reliability of non-landslide samples, which will pose a serious risk of including potential landslides and lead to erroneous outcomes in training data. Furthermore, diverse machine-learning models exhibit distinct classification capabilities, and applying a single model can readily result in over-fitting of the dataset and introduce potential uncertainties in predictions. To address these problems, taking Chenxi County, a hilly and mountainous area in southern China, as an example, this research proposes a strategy-coupling optimised sampling with heterogeneous ensemble machine learning to enhance the accuracy of landslide susceptibility prediction. Initially, 21 landslide impact factors were derived from five aspects: geology, hydrology, topography, meteorology, human activities, and geographical environment. Then, these factors were screened through a correlation analysis and collinearity diagnosis. Afterwards, an optimised sampling (OS) method was utilised to select negative samples by fusing the reliability of non-landslide samples and certainty factor values on the basis of the environmental similarity and statistical model. Subsequently, the adopted non-landslide samples and historical landslides were combined to create machine-learning datasets. Finally, baseline models (support vector machine, random forest, and back propagation neural network) and the stacking ensemble model were employed to predict susceptibility. The findings indicated that the OS method, considering the reliability of non-landslide samples, achieved higher-quality negative samples than currently widely used sampling methods. The stacking ensemble machine-learning model outperformed those three baseline models. Notably, the accuracy of the hybrid OS–Stacking model is most promising, up to 97.1%. The integrated strategy significantly improves the prediction of landslide susceptibility and makes it reliable and effective for assessing regional geohazard risk. Full article
Show Figures

Figure 1

11 pages, 12811 KiB  
Article
On the Initial Fabric of Naturally Occurring and Reconstituted Weakly Cemented Geomaterials
by Mohd Ilyas Bhat, Bhupendra Chand and Tejas Gorur Murthy
Minerals 2024, 14(10), 1000; https://doi.org/10.3390/min14101000 - 30 Sep 2024
Viewed by 350
Abstract
The understanding of naturally occurring materials such as clay, sand, hard and soft rocks under a common theoretical framework has been a topic of persistent research interest. Over the past few decades, various sample reconstitution techniques have been developed in the literature to [...] Read more.
The understanding of naturally occurring materials such as clay, sand, hard and soft rocks under a common theoretical framework has been a topic of persistent research interest. Over the past few decades, various sample reconstitution techniques have been developed in the literature to mimic in situ conditions, and to parse carefully the influence of various components in a cohesive-frictional geomaterial such that their behavior can be folded into the broad ambit of a continuum mechanics framework. The initial fabric of natural rock specimens is compared with reconstituted cemented sand samples using X-ray computed tomography (XRCT) scans. The efficacy of laboratory reconstitution techniques in replicating the initial microstructural features of natural rocks is evaluated here. Additionally, discrete element method (DEM) protocols which are often employed in generating cohesive granular ensembles are employed here and compared against the naturally occurring and artificially reconstituted fabric. A significant difference is observed in the grain boundaries of reconstituted and naturally occurring rocks. Additionally, the arrangement of particles, the orientation of grain contacts, and their coordination number are examined to assess the efficacy of laboratory-reconstituted specimens at micro-length scale. Full article
Show Figures

Figure 1

29 pages, 4571 KiB  
Article
Natural Language Inference with Transformer Ensembles and Explainability Techniques
by Isidoros Perikos and Spyro Souli
Electronics 2024, 13(19), 3876; https://doi.org/10.3390/electronics13193876 - 30 Sep 2024
Viewed by 445
Abstract
Natural language inference (NLI) is a fundamental and quite challenging task in natural language processing, requiring efficient methods that are able to determine whether given hypotheses derive from given premises. In this paper, we apply explainability techniques to natural-language-inference methods as a means [...] Read more.
Natural language inference (NLI) is a fundamental and quite challenging task in natural language processing, requiring efficient methods that are able to determine whether given hypotheses derive from given premises. In this paper, we apply explainability techniques to natural-language-inference methods as a means to illustrate the decision-making procedure of its methods. First, we investigate the performance and generalization capabilities of several transformer-based models, including BERT, ALBERT, RoBERTa, and DeBERTa, across widely used datasets like SNLI, GLUE Benchmark, and ANLI. Then, we employ stacking-ensemble techniques to leverage the strengths of multiple models and improve inference performance. Experimental results demonstrate significant improvements of the ensemble models in inference tasks, highlighting the effectiveness of stacking. Specifically, our best-performing ensemble models surpassed the best-performing individual transformer by 5.31% in accuracy on MNLI-m and MNLI-mm tasks. After that, we implement LIME and SHAP explainability techniques to shed light on the decision-making of the transformer models, indicating how specific words and contextual information are utilized in the transformer inferences procedures. The results indicate that the model properly leverages contextual information and individual words to make decisions but, in some cases, find difficulties in inference scenarios with metaphorical connections which require deeper inferential reasoning. Full article
(This article belongs to the Special Issue Advances in Artificial Intelligence Engineering)
Show Figures

Figure 1

Back to TopTop