Innovative Approaches To Enhance Data Science Optimization
Innovative Approaches To Enhance Data Science Optimization
ISSN No:-2456-2165
T. Gopi Krishna3
3
Department of Computer Science & Engineering,
School of Electrical Engineering and Computing,Adama Science & Technology University, Ethiopia
In our study, we utilized a machine learning model to Arabic Text corpus and manually organized the dialect words
evaluate the performance of the system in the preprocessing during this phase.
stage. A training set was created by compiling a dataset of
described Arabic dialects. The corpus of the training dataset A. Dataset
includes several dialects, as detailed in Table-5 (Libya-1, We've conducted preprocessing on a moderately sized
Morocco-2, Egypt-3, Jordan-4, Palestine-5, and Sudan-6). dataset of Arabic dialects, specifically aligned with the
Notably, our simple training model produced well-optimized Modern Standard Arabic Language. Our model was
results for the proposed framework. To assess the model's constructed using a machine-learning approach, building
reliability, we intentionally selected a small subset from the upon the foundation of a developed model for the dataset [9,
10]. Table 4. shows transformations.
VI. ANALYSIS OF RESULTS transformation adapts the rules into conditional rules,
aligning with the principles of soft set theory. This
Our implementation of chosen methods aimed to straightforward and versatile approach ensures that the data is
optimize the functionality of our machine learning model. appropriately prepared for subsequent use. The epoch, a
The data presented in Table-6 outlines the contents of our crucial phase in training, utilizes all available information to
dataset, which encompasses a variety of Arabic-language refine parameters and enhance accuracy during testing.
documents categorized across different topics. Additionally, Table-6 provides a visual representation of the numerical
we conducted model training using a basic rules-based table. values employed to instruct optimization techniques within
This training process facilitated the conversion of the model's the suggested model.
rules into a binary table format, representing the soft set. This
Table 5: Training Data Results Using Various Optimization Methods In The Proposed Model
SNo Iteration Method-1 M-2 M-3 M-4 M-5 M-6
Progress Data Scaling & Feature Feature Data Parallel
Cleaning Normalization Selection Engineering Augmentation Processing
1 00 0.00 0.00 0.00 0.00 0.00 0.00
2 20 0.893 0.923 0.881 0.883 0.876 0.832
3 40 0.899 0.926 0.871 0.920 0.894 0.836
4 60 0.901 0.944 0.912 0.927 0.900 0.921
5 80 0.924 0.968 0.913 0.936 0.922 0.951
6 100 0.941 0.944 0.955 0.957 0.958 0.961
Table 6. illustrates the effectiveness of our suggested for the first 100 rounds, showcasing improved performance at
methods, revealing a favorable trend around the 60th epoch, each 20th epoch, resulting in heightened accuracies through
where the loss level stabilizes. The model underwent training our optimized approaches [13, 15].