Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,291)

Search Parameters:
Keywords = multi-task learning

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
15 pages, 2347 KiB  
Article
An Underwater Multi-Label Classification Algorithm Based on a Bilayer Graph Convolution Learning Network with Constrained Codec
by Yun Li, Su Wang, Jiawei Mo and Xin Wei
Electronics 2024, 13(16), 3134; https://doi.org/10.3390/electronics13163134 (registering DOI) - 7 Aug 2024
Abstract
Within the domain of multi-label classification for micro-videos, utilizing terrestrial datasets as a foundation, researchers have embarked on profound endeavors yielding extraordinary accomplishments. The research into multi-label classification based on underwater micro-video datasets is still in the preliminary stage. There are some challenges: [...] Read more.
Within the domain of multi-label classification for micro-videos, utilizing terrestrial datasets as a foundation, researchers have embarked on profound endeavors yielding extraordinary accomplishments. The research into multi-label classification based on underwater micro-video datasets is still in the preliminary stage. There are some challenges: the severe color distortion and visual blurring in underwater visual imaging due to water molecular scattering and absorption, the difficulty in acquiring underwater short video datasets, the sparsity of underwater short video modality features, and the formidable task of achieving high-precision underwater multi-label classification. To address these issues, a bilayer graph convolution learning network based on constrained codec (BGCLN) is established in this paper. Specifically, modality-common representation is constructed to complete the representation of common information and specific information based on the constrained codec network. Then, the attention-driven double-layer graph convolutional network module is designed to mine the correlation information between labels and enhance the modality representation. Finally, the combined modality representation fusion and multi-label classification module are used to obtain the category classifier prediction. In the underwater video multi-label classification dataset (UVMCD), the effectiveness and high classification accuracy of the proposed BGCLN have been proved by numerous experiments. Full article
(This article belongs to the Special Issue Underwater Robotics: Theory, Methods and Applications)
Show Figures

Figure 1

16 pages, 9003 KiB  
Article
SiM-YOLO: A Wood Surface Defect Detection Method Based on the Improved YOLOv8
by Honglei Xi, Rijun Wang, Fulong Liang, Yesheng Chen, Guanghao Zhang and Bo Wang
Coatings 2024, 14(8), 1001; https://doi.org/10.3390/coatings14081001 (registering DOI) - 7 Aug 2024
Abstract
Wood surface defect detection is a challenging task due to the complexity and variability of defect types. To address these challenges, this paper introduces a novel deep learning approach named SiM-YOLO, which is built upon the YOLOv8 object detection framework. A fine-grained convolutional [...] Read more.
Wood surface defect detection is a challenging task due to the complexity and variability of defect types. To address these challenges, this paper introduces a novel deep learning approach named SiM-YOLO, which is built upon the YOLOv8 object detection framework. A fine-grained convolutional structure, SPD-Conv, is introduced with the aim of preserving detailed defect information during the feature extraction process, thus enabling the model to capture the subtle variations and complex details of wood surface defects. In the feature fusion stage, a SiAFF-PANet-based wood defect feature fusion module is designed to improve the model’s ability to focus on local contextual information and enhance defect localization. For classification and regression tasks, the multi-attention detection head (MADH) is employed to capture cross-channel information and the accurate spatial localization of defects. In addition, MPDIoU is employed to optimize the loss function of the model to reduce the leakage of detection due to defect overlap. The experimental results show that SiM-YOLO achieves superior performance compared to the state-of-the-art YOLO algorithm, with a 9.3% improvement in mAP over YOLOX and a 4.3% improvement in mAP over YOLOv8. The Grad-CAM visualization further illustrates that SiM-YOLO provides more accurate defect localization and effectively reduces misdetection and omission issues. This study highlights the effectiveness of SiM-YOLO for wood surface defect detection and offers valuable insights for future research and practical applications in quality control. Full article
(This article belongs to the Section Surface Characterization, Deposition and Modification)
Show Figures

Figure 1

19 pages, 4475 KiB  
Article
A Multi-Level Cross-Attention Image Registration Method for Visible and Infrared Small Unmanned Aerial Vehicle Targets via Image Style Transfer
by Wen Jiang, Hanxin Pan, Yanping Wang, Yang Li, Yun Lin and Fukun Bi
Remote Sens. 2024, 16(16), 2880; https://doi.org/10.3390/rs16162880 - 7 Aug 2024
Abstract
Small UAV target detection and tracking based on cross-modality image fusion have gained widespread attention. Due to the limited feature information available from small UAVs in images, where they occupy a minimal number of pixels, the precision required for detection and tracking algorithms [...] Read more.
Small UAV target detection and tracking based on cross-modality image fusion have gained widespread attention. Due to the limited feature information available from small UAVs in images, where they occupy a minimal number of pixels, the precision required for detection and tracking algorithms is particularly high in complex backgrounds. Image fusion techniques can enrich the detailed information for small UAVs, showing significant advantages under extreme lighting conditions. Image registration is a fundamental step preceding image fusion. It is essential to achieve accurate image alignment before proceeding with image fusion to prevent severe ghosting and artifacts. This paper specifically focused on the alignment of small UAV targets within infrared and visible light imagery. To address this issue, this paper proposed a cross-modality image registration network based on deep learning, which includes a structure preservation and style transformation network (SPSTN) and a multi-level cross-attention residual registration network (MCARN). Firstly, the SPSTN is employed for modality transformation, transferring the cross-modality task into a single-modality task to reduce the information discrepancy between modalities. Then, the MCARN is utilized for single-modality image registration, capable of deeply extracting and fusing features from pseudo infrared and visible images to achieve efficient registration. To validate the effectiveness of the proposed method, comprehensive experimental evaluations were conducted on the Anti-UAV dataset. The extensive evaluation results validate the superiority and universality of the cross-modality image registration framework proposed in this paper, which plays a crucial role in subsequent image fusion tasks for more effective target detection. Full article
(This article belongs to the Special Issue Deep Learning and Computer Vision in Remote Sensing-III)
Show Figures

Figure 1

23 pages, 14390 KiB  
Article
Multi-Task Intelligent Monitoring of Construction Safety Based on Computer Vision
by Lingfeng Liu, Zhigang Guo, Zhengxiong Liu, Yaolin Zhang, Ruying Cai, Xin Hu, Ran Yang and Gang Wang
Buildings 2024, 14(8), 2429; https://doi.org/10.3390/buildings14082429 - 6 Aug 2024
Viewed by 208
Abstract
Effective safety management is vital for ensuring construction safety. Traditional safety inspections in construction heavily rely on manual labor, which is both time-consuming and labor-intensive. Extensive research has been conducted integrating computer-vision technologies to facilitate intelligent surveillance and improve safety measures. However, existing [...] Read more.
Effective safety management is vital for ensuring construction safety. Traditional safety inspections in construction heavily rely on manual labor, which is both time-consuming and labor-intensive. Extensive research has been conducted integrating computer-vision technologies to facilitate intelligent surveillance and improve safety measures. However, existing research predominantly focuses on singular tasks, while construction environments necessitate comprehensive analysis. This study introduces a multi-task computer vision technology approach for the enhanced monitoring of construction safety. The process begins with the collection and processing of multi-source video surveillance data. Subsequently, YOLOv8, a deep learning-based computer vision model, is adapted to meet specific task requirements by modifying the head component of the framework. This adaptation enables efficient detection and segmentation of construction elements, as well as the estimation of person and machine poses. Moreover, a tracking algorithm integrates these capabilities to continuously monitor detected elements, thereby facilitating the proactive identification of unsafe practices on construction sites. This paper also presents a novel Integrated Excavator Pose (IEP) dataset designed to address the common challenges associated with different single datasets, thereby ensuring accurate detection and robust application in practical scenarios. Full article
(This article belongs to the Special Issue Intelligence and Automation in Construction Industry)
Show Figures

Figure 1

26 pages, 6549 KiB  
Article
Reinforcement-Learning-Based Multi-UAV Cooperative Search for Moving Targets in 3D Scenarios
by Yifei Liu, Xiaoshuai Li, Jian Wang, Feiyu Wei and Junan Yang
Viewed by 224
Abstract
Most existing multi-UAV collaborative search methods only consider scenarios of two-dimensional path planning or static target search. To be close to the practical scenario, this paper proposes a path planning method based on an action-mask-based multi-agent proximal policy optimization (AM-MAPPO) algorithm for multiple [...] Read more.
Most existing multi-UAV collaborative search methods only consider scenarios of two-dimensional path planning or static target search. To be close to the practical scenario, this paper proposes a path planning method based on an action-mask-based multi-agent proximal policy optimization (AM-MAPPO) algorithm for multiple UAVs searching for moving targets in three-dimensional (3D) environments. In particular, a multi-UAV high–low altitude collaborative search architecture is introduced that not only takes into account the extensive detection range of high-altitude UAVs but also leverages the benefit of the superior detection quality of low-altitude UAVs. The optimization objective of the search task is to minimize the uncertainty of the search area while maximizing the number of captured moving targets. The path planning problem for moving target search in a 3D environment is formulated and addressed using the AM-MAPPO algorithm. The proposed method incorporates a state representation mechanism based on field-of-view encoding to handle dynamic changes in neural network input dimensions and develops a rule-based target capture mechanism and an action-mask-based collision avoidance mechanism to enhance the AM-MAPPO algorithm’s convergence speed. Experimental results demonstrate that the proposed algorithm significantly reduces regional uncertainty and increases the number of captured moving targets compared to other deep reinforcement learning methods. Ablation studies further indicate that the proposed action mask mechanism, target capture mechanism, and collision avoidance mechanism of the AM-MAPPO algorithm can improve the algorithm’s effectiveness, target capture capability, and UAVs’ safety, respectively. Full article
Show Figures

Figure 1

16 pages, 1818 KiB  
Article
FFA-BiGRU: Attention-Based Spatial-Temporal Feature Extraction Model for Music Emotion Classification
by Yuping Su, Jie Chen, Ruiting Chai, Xiaojun Wu and Yumei Zhang
Appl. Sci. 2024, 14(16), 6866; https://doi.org/10.3390/app14166866 - 6 Aug 2024
Viewed by 193
Abstract
Music emotion recognition is becoming an important research direction due to its great significance for music information retrieval, music recommendation, and so on. In the task of music emotion recognition, the key to achieving accurate emotion recognition lies in how to extract the [...] Read more.
Music emotion recognition is becoming an important research direction due to its great significance for music information retrieval, music recommendation, and so on. In the task of music emotion recognition, the key to achieving accurate emotion recognition lies in how to extract the affect-salient features fully. In this paper, we propose an end-to-end spatial-temporal feature extraction method named FFA-BiGRU for music emotion classification. Taking the log Mel-spectrogram of music audio as the input, this method employs an attention-based convolutional residual module named FFA, which serves as a spatial feature learning module to obtain multi-scale spatial features. In the FFA module, three group architecture blocks extract multi-level spatial features, each of which consists of a stack of multiple channel-spatial attention-based residual blocks. Then, the output features from FFA are fed into the bidirectional gated recurrent units (BiGRU) module to capture the temporal features of music further. In order to make full use of the extracted spatial and temporal features, the output feature maps of FFA and those of the BiGRU are concatenated in the channel dimension. Finally, the concatenated features are passed through fully connected layers to predict the emotion classification results. The experimental results of the EMOPIA dataset show that the proposed model achieves better classification accuracy than the existing baselines. Meanwhile, the ablation experiments also demonstrate the effectiveness of each part of the proposed method. Full article
Show Figures

Figure 1

17 pages, 6217 KiB  
Article
Machine-Learning-Enabled Diagnostics with Improved Visualization of Disease Lesions in Chest X-ray Images
by Md Fashiar Rahman, Tzu-Liang (Bill) Tseng, Michael Pokojovy, Peter McCaffrey, Eric Walser, Scott Moen, Alex Vo and Johnny C. Ho
Diagnostics 2024, 14(16), 1699; https://doi.org/10.3390/diagnostics14161699 - 6 Aug 2024
Viewed by 257
Abstract
The class activation map (CAM) represents the neural-network-derived region of interest, which can help clarify the mechanism of the convolutional neural network’s determination of any class of interest. In medical imaging, it can help medical practitioners diagnose diseases like COVID-19 or pneumonia by [...] Read more.
The class activation map (CAM) represents the neural-network-derived region of interest, which can help clarify the mechanism of the convolutional neural network’s determination of any class of interest. In medical imaging, it can help medical practitioners diagnose diseases like COVID-19 or pneumonia by highlighting the suspicious regions in Computational Tomography (CT) or chest X-ray (CXR) film. Many contemporary deep learning techniques only focus on COVID-19 classification tasks using CXRs, while few attempt to make it explainable with a saliency map. To fill this research gap, we first propose a VGG-16-architecture-based deep learning approach in combination with image enhancement, segmentation-based region of interest (ROI) cropping, and data augmentation steps to enhance classification accuracy. Later, a multi-layer Gradient CAM (ML-Grad-CAM) algorithm is integrated to generate a class-specific saliency map for improved visualization in CXR images. We also define and calculate a Severity Assessment Index (SAI) from the saliency map to quantitatively measure infection severity. The trained model achieved an accuracy score of 96.44% for the three-class CXR classification task, i.e., COVID-19, pneumonia, and normal (healthy patients), outperforming many existing techniques in the literature. The saliency maps generated from the proposed ML-GRAD-CAM algorithm are compared with the original Gran-CAM algorithm. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

13 pages, 441 KiB  
Article
Cooperative MARL-PPO Approach for Automated Highway Platoon Merging
by Máté Kolat and Tamás Bécsi
Electronics 2024, 13(15), 3102; https://doi.org/10.3390/electronics13153102 - 5 Aug 2024
Viewed by 292
Abstract
This paper presents a cooperative highway platooning strategy that integrates Multi-Agent Reinforcement Learning (MARL) with Proximal Policy Optimization (PPO) to effectively manage the complex task of merging. In modern transportation systems, platooning—where multiple vehicles travel closely together under coordinated control—promises significant improvements in [...] Read more.
This paper presents a cooperative highway platooning strategy that integrates Multi-Agent Reinforcement Learning (MARL) with Proximal Policy Optimization (PPO) to effectively manage the complex task of merging. In modern transportation systems, platooning—where multiple vehicles travel closely together under coordinated control—promises significant improvements in traffic flow and fuel efficiency. However, the challenge of merging, which involves dynamically adjusting the formation to incorporate new vehicles, remains challenging. Our approach leverages the strengths of MARL to enable individual vehicles within a platoon to learn optimal behaviors through interactions. PPO ensures stable and efficient learning by optimizing policies balancing exploration and exploitation. Simulation results show that our method achieves merging with safety and operational efficiency. Full article
(This article belongs to the Special Issue New Insights in Multi-Agent Systems and Intelligent Control)
Show Figures

Figure 1

23 pages, 8343 KiB  
Article
A Multi-Area Task Path-Planning Algorithm for Agricultural Drones Based on Improved Double Deep Q-Learning Net
by Jian Li, Weijian Zhang, Junfeng Ren, Weilin Yu, Guowei Wang, Peng Ding, Jiawei Wang and Xuen Zhang
Agriculture 2024, 14(8), 1294; https://doi.org/10.3390/agriculture14081294 - 5 Aug 2024
Viewed by 297
Abstract
With the global population growth and increasing food demand, the development of precision agriculture has become particularly critical. In precision agriculture, accurately identifying areas of nitrogen stress in crops and planning precise fertilization paths are crucial. However, traditional coverage path-planning (CPP) typically considers [...] Read more.
With the global population growth and increasing food demand, the development of precision agriculture has become particularly critical. In precision agriculture, accurately identifying areas of nitrogen stress in crops and planning precise fertilization paths are crucial. However, traditional coverage path-planning (CPP) typically considers only single-area tasks and overlooks the multi-area tasks CPP. To address this problem, this study proposed a Regional Framework for Coverage Path-Planning for Precision Fertilization (RFCPPF) for crop protection UAVs in multi-area tasks. This framework includes three modules: nitrogen stress spatial distribution extraction, multi-area tasks environmental map construction, and coverage path-planning. Firstly, Sentinel-2 remote-sensing images are processed using the Google Earth Engine (GEE) platform, and the Green Normalized Difference Vegetation Index (GNDVI) is calculated to extract the spatial distribution of nitrogen stress. A multi-area tasks environmental map is constructed to guide multiple UAV agents. Subsequently, improvements based on the Double Deep Q Network (DDQN) are introduced, incorporating Long Short-Term Memory (LSTM) and dueling network structures. Additionally, a multi-objective reward function and a state and action selection strategy suitable for stress area plant protection operations are designed. Simulation experiments verify the superiority of the proposed method in reducing redundant paths and improving coverage efficiency. The proposed improved DDQN achieved an overall step count that is 60.71% of MLP-DDQN and 90.55% of Breadth-First Search–Boustrophedon Algorithm (BFS-BA). Additionally, the total repeated coverage rate was reduced by 7.06% compared to MLP-DDQN and by 8.82% compared to BFS-BA. Full article
(This article belongs to the Section Digital Agriculture)
Show Figures

Figure 1

16 pages, 4635 KiB  
Article
Deep Learning-Based Defects Detection in Keyhole TIG Welding with Enhanced Vision
by Xuan Zhang, Shengbin Zhao and Mingdi Wang
Materials 2024, 17(15), 3871; https://doi.org/10.3390/ma17153871 - 5 Aug 2024
Viewed by 303
Abstract
Keyhole tungsten inert gas (keyhole TIG) welding is renowned for its advanced efficiency, necessitating a real-time defect detection method that integrates deep learning and enhanced vision techniques. This study employs a multi-layer deep neural network trained on an extensive welding image dataset. Neural [...] Read more.
Keyhole tungsten inert gas (keyhole TIG) welding is renowned for its advanced efficiency, necessitating a real-time defect detection method that integrates deep learning and enhanced vision techniques. This study employs a multi-layer deep neural network trained on an extensive welding image dataset. Neural networks can capture complex nonlinear relationships through multi-layer transformations without manual feature selection. Conversely, the nonlinear modeling ability of support vector machines (SVM) is limited by manually selected kernel functions and parameters, resulting in poor performance for recognizing burn-through and good welds images. SVMs handle only lower-level features such as porosity and excel only in detecting simple edges and shapes. However, neural networks excel in processing deep feature maps of “molten pools” and can encode deep defects that are often confused in keyhole TIG. Applying a four-class classification task to weld pool images, the neural network adeptly distinguishes various weld states, including good welds, burn-through, partial penetration, and undercut. Experimental results demonstrate high accuracy and real-time performance. A comprehensive dataset, prepared through meticulous preprocessing and augmentation, ensures reliable results. This method provides an effective solution for quality control and defect prevention in keyhole TIG welding process. Full article
(This article belongs to the Special Issue Advanced Welding in Alloys and Composites)
Show Figures

Figure 1

17 pages, 1054 KiB  
Article
Integration of Relation Filtering and Multi-Task Learning in GlobalPointer for Entity and Relation Extraction
by Bin Liu, Jialin Tao, Wanyuan Chen, Yijie Zhang, Min Chen, Lei He and Dan Tang
Appl. Sci. 2024, 14(15), 6832; https://doi.org/10.3390/app14156832 - 5 Aug 2024
Viewed by 336
Abstract
The rise of knowledge graphs has been instrumental in advancing artificial intelligence (AI) research. Extracting entity and relation triples from unstructured text is crucial for the construction of knowledge graphs. However, Chinese text has a complex grammatical structure, which may lead to the [...] Read more.
The rise of knowledge graphs has been instrumental in advancing artificial intelligence (AI) research. Extracting entity and relation triples from unstructured text is crucial for the construction of knowledge graphs. However, Chinese text has a complex grammatical structure, which may lead to the problem of overlapping entities. Previous pipeline models have struggled to address such overlap problems effectively, while joint models require entity annotations for each predefined relation in the set, which results in redundant relations. In addition, the traditional models often lead to task imbalance by overlooking the differences between tasks. To tackle these challenges, this research proposes a global pointer network based on relation prediction and loss function improvement (GPRL) for joint extraction of entities and relations. Experimental evaluations on the publicly available Chinese datasets DuIE2.0 and CMeIE demonstrate that the GPRL model achieves a 1.2–26.1% improvement in F1 score compared with baseline models. Further, experiments of overlapping classification conducted on CMeIE have also verified the effectiveness of overlapping triad extraction and ablation experiments. The model is helpful in identifying entities and relations accurately and can reduce redundancy by leveraging relation filtering and the global pointer network. In addition, the incorporation of a multi-task learning framework balances the loss functions of multiple tasks and enhances task interactions. Full article
Show Figures

Figure 1

30 pages, 1329 KiB  
Review
Understanding and Therapeutic Application of Immune Response in Major Histocompatibility Complex (MHC) Diversity Using Multimodal Artificial Intelligence
by Yasunari Matsuzaka and Ryu Yashiro
BioMedInformatics 2024, 4(3), 1835-1864; https://doi.org/10.3390/biomedinformatics4030101 - 5 Aug 2024
Viewed by 314
Abstract
Human Leukocyte Antigen (HLA) is like a device that monitors the internal environment of the body. T lymphocytes immediately recognize the HLA molecules that are expressed on the surface of the cells of the different individual, attacking it defeats microorganisms that is one [...] Read more.
Human Leukocyte Antigen (HLA) is like a device that monitors the internal environment of the body. T lymphocytes immediately recognize the HLA molecules that are expressed on the surface of the cells of the different individual, attacking it defeats microorganisms that is one of the causes of rejection in organ transplants performed between people with unmatched HLA types. Over 2850 and 3580 different polymorphisms have been reported for HLA-A and HLA-B respectively, around the world. HLA genes are associated with the risk of developing a variety of diseases, including autoimmune diseases, and play an important role in pathological conditions. By using a deep learning method called multi-task learning to simultaneously predict the gene sequences of multiple HLA genes, it is possible to improve accuracy and shorten execution time. Some new systems use a model called convolutional neural network (CNNs) in deep learning, which uses neural networks consisting of many layers and can learn complex correlations between SNP information and HLA gene sequences based on reference data for HLA imputation, which serves as training data. The learned model can output predicted values of HLA gene sequences with high accuracy using SNP information as input. To investigate which part of the input information surrounding the HLA gene is used to make learning predictions, predictions were made using not only a small number of nearby SNP information but also many SNP information distributed over a wider area by visualizing the learning information of the model. While conventional methods are strong at learning using nearly SNP information and not good at learning using SNP information located at distant locations, some new systems are thought that prediction accuracy may have improved because this problem was overcome. HLA genes are involved in the onset of a variety of diseases and are attracting attention. As an important area from the perspective of elucidating pathological conditions and realizing personalized medicine. The applied multi-task learning to two different HLA imputation reference panels—a Japanese panel (n = 1118) and type I diabetes genetics consortium panel (n = 5122). Through 10-fold cross-validation on these panels, the multi-task learning achieved higher imputation accuracy than conventional methods, especially for imputing low-frequency and rare HLA alleles. The increased prediction accuracy of HLA gene sequences is expected to increase the reliability of HLA analysis, including integrated analysis between different racial populations, and is expected to greatly contribute to the identification of HLA gene sequences associated with diseases and further elucidation of pathological conditions. Full article
(This article belongs to the Special Issue Feature Papers on Methods in Biomedical Informatics)
Show Figures

Figure 1

25 pages, 636 KiB  
Article
A User-Centered Framework for Data Privacy Protection Using Large Language Models and Attention Mechanisms
by Shutian Zhou, Zizhe Zhou, Chenxi Wang, Yuzhe Liang, Liangyu Wang, Jiahe Zhang, Jinming Zhang and Chunli Lv
Appl. Sci. 2024, 14(15), 6824; https://doi.org/10.3390/app14156824 - 5 Aug 2024
Viewed by 333
Abstract
This paper introduces a user-centered data privacy protection framework utilizing large language models (LLMs) and user attention mechanisms, which are tailored to address urgent privacy concerns in sensitive data processing domains like financial computing and facial recognition. The innovation lies in a novel [...] Read more.
This paper introduces a user-centered data privacy protection framework utilizing large language models (LLMs) and user attention mechanisms, which are tailored to address urgent privacy concerns in sensitive data processing domains like financial computing and facial recognition. The innovation lies in a novel user attention mechanism that dynamically adjusts attention weights based on data characteristics and user privacy needs, enhancing the ability to identify and protect sensitive information effectively. Significant methodological advancements differentiate our approach from existing techniques by incorporating user-specific attention into traditional LLMs, ensuring both data accuracy and privacy. We succinctly highlight the enhanced performance of this framework through a selective presentation of experimental results across various applications. Notably, in computer vision, the application of our user attention mechanism led to improved metrics over traditional multi-head and self-attention methods: FasterRCNN models achieved precision, recall, and accuracy rates of 0.82, 0.79, and 0.80, respectively. Similar enhancements were observed with SSD, YOLO, and EfficientDet models with notable increases in all performance metrics. In natural language processing tasks, our framework significantly boosted the performance of models like Transformer, BERT, CLIP, BLIP, and BLIP2, demonstrating the framework’s adaptability and effectiveness. These streamlined results underscore the practical impact and the technological advancement of our proposed framework, confirming its superiority in enhancing privacy protection without compromising on data processing efficacy. Full article
(This article belongs to the Special Issue Cloud Computing: Privacy Protection and Data Security)
Show Figures

Figure 1

19 pages, 1076 KiB  
Article
TRUST-ME: Trust-Based Resource Allocation and Server Selection in Multi-Access Edge Computing
by Sean Tsikteris, Aisha B Rahman, Md. Sadman Siraj and Eirini Eleni Tsiropoulou
Future Internet 2024, 16(8), 278; https://doi.org/10.3390/fi16080278 - 4 Aug 2024
Viewed by 305
Abstract
Multi-access edge computing (MEC) has attracted the interest of the research and industrial community to support Internet of things (IoT) applications by enabling efficient data processing and minimizing latency. This paper presents significant contributions toward optimizing the resource allocation and enhancing the decision-making [...] Read more.
Multi-access edge computing (MEC) has attracted the interest of the research and industrial community to support Internet of things (IoT) applications by enabling efficient data processing and minimizing latency. This paper presents significant contributions toward optimizing the resource allocation and enhancing the decision-making process in edge computing environments. Specifically, the TRUST-ME model is introduced, which consists of multiple edge servers and IoT devices, i.e., users, with varied computing tasks offloaded to the MEC servers. A utility function was designed to quantify the benefits in terms of latency and cost for the IoT device while utilizing the MEC servers’ computing capacities. The core innovation of our work is a novel trust model that was designed to evaluate the IoT devices’ confidence in MEC servers. This model integrates both direct and indirect trust and reflects the trustworthiness of the servers based on the direct interactions and social feedback from other devices using the same servers. This dual trust approach helps with accurately gauging the reliability of MEC services and ensuring more informed decision making. A reinforcement learning framework based on the optimistic Q-learning with an upper confidence bounds action selection algorithm enables the IoT devices to autonomously select a MEC server to process their computing tasks. Also, a multilateral bargaining model is proposed for fair resource allocation of the MEC servers’ computing resources to the users while accounting for their computing demands. Numerical simulations demonstrated the operational effectiveness, convergence, and scalability of the TRUST-ME model, which was validated through real-world scenarios and comprehensive comparative evaluations against existing approaches. Full article
Show Figures

Figure 1

14 pages, 605 KiB  
Article
A Hierarchical Multi-Task Learning Framework for Semantic Annotation in Tabular Data
by Jie Wu and Mengshu Hou
Entropy 2024, 26(8), 664; https://doi.org/10.3390/e26080664 - 4 Aug 2024
Viewed by 298
Abstract
To optimize the utilization and analysis of tables, it is essential to recognize and understand their semantics comprehensively. This requirement is especially critical given that many tables lack explicit annotations, necessitating the identification of column types and inter-column relationships. Such identification can significantly [...] Read more.
To optimize the utilization and analysis of tables, it is essential to recognize and understand their semantics comprehensively. This requirement is especially critical given that many tables lack explicit annotations, necessitating the identification of column types and inter-column relationships. Such identification can significantly augment data quality, streamline data integration, and support data analysis and mining. Current table annotation models often address each subtask independently, which may result in the neglect of constraints and contextual information, causing relational ambiguities and inference errors. To address this issue, we propose a unified multi-task learning framework capable of concurrently handling multiple tasks within a single model, including column named entity recognition, column type identification, and inter-column relationship detection. By integrating these tasks, the framework exploits their interrelations, facilitating the exchange of shallow features and the sharing of representations. Their cooperation enables each task to leverage insights from the others, thereby improving the performance of individual subtasks and enhancing the model’s overall generalization capabilities. Notably, our model is designed to employ only the internal information of tabular data, avoiding reliance on external context or knowledge graphs. This design ensures robust performance even with limited input information. Extensive experiments demonstrate the superior performance of our model across various tasks, validating the effectiveness of unified multi-task learning framework in the recognition and comprehension of table semantics. Full article
(This article belongs to the Special Issue Natural Language Processing and Data Mining)
Show Figures

Figure 1

Back to TopTop