Chapter 1
Chapter 1
1.1 Background
Based on the advanced digital technology and communication networks,
humans are more conscious in obtaining enhanced health information along with EHR
[1]. The highly advanced hospitals with specialized equipment are rapidly increasing
worldwide. Hence, patients pay more attention in visiting these medical care service
providers and, they are interested in consulting popular doctors for better disease
cal, pharmacy, personal and account
ting data are needed to be stored properly in segmented way. As days goes on, the
amount of data stored has grown rapidly. Hence, there is a need for upgrading the old
storage system, which is considered as a complicate process, especially in security
issues [2]. Also, there is a need for interchanging of information regarding health issues
between the providers in the healthcare environment when the patients prefer various
hospitals for treatments. This is considered as a complicated process. Hence, there
exists an urge to covert the health in formations into digital form along with proper
network for communication for sharing the generations regarding to the health issues
of patients. A consolidated healthcare data management is required among the doctors,
patients, hospitals, pharmaceutical manufactures, insurance firms and medical labs,
which lie at several geographical locations [3]. For obtaining a solution for the above-
mentioned complications, a new system is required to be implemented with efficient
storage space, safely stored EHR and a properly validated retrieval mechanism. Such
-
sharing and patient-provider relationship with more relevant to patient-centric data
management in health care environment.
In day-to-day medical platforms, there exists a rapid development in medical
imaging and detection techniques which have become most important element in
disease detection. Medical images embrace a significant role in storing information of
nerves, brain, heart etc. for obtaining the interior vision of human body [4]. The medical
images consist of enormous mathematical applications to analyze, whether the health
tissue is affected or not. Although, missing of any specific data or area during the
process of medical imaging may lead to death. Hence, this is considered as a main
complication in medical imaging that is to utilize the images without any kind of
1
information or data loss. It is highly feasible and during the process of acquiring of
medical images, the corruption of network by artifacts and noise may takes place. The
term noise is defined as changing of original pixel values of images, randomly. Due to
this interruption of noises, the quality of the images is reduced and is highly remarkable
when the images comprise of small size and lower contrast in comparison with the
images obtained from the conventional image processing [5]. Damaging the medical
image noises is severely critical because these noises will minimize the image quality
and also confuse the disease detection processing. Hence, the process of de-noising in
medical images is unavoidable and is considered as essential pre-processing stage in
the medical imaging systems.
The examining of the medical image model is generally carried out for
identification of single disease. But, the demand for the detection of multi-level diseases
in medical applications and the multi-function solutions are needed to be explored.
Many health care analysis systems consider only one disease at a time [8]. Several
articles are also focused on one specific disease. When, the organizations need to look
after the reports of their patients, they must establish many models, but the existing
process is utilized for analyzing only a specific disease [7]. The death rate is increasing
now-a-days due to the improper identification of the specific disease exactly. The cured
patient from one disease may tend to affect from another disease due to this improper
detection [9]. Some of the available systems utilize several parameters during the
process of analysis of disease. Due to this, it is not possible to identify one disease,
which is caused by the side effect of another disease [6]. For example, due to diabetes,
there exists the chance of neuropath, heart disease, hearing loss, dementia and
retinopathy. Hence, there is a need for multi-disease model.
2
clinical use, administrative purposes, and surveillance. EHR is important for
Data Data
collection preprocessing
EHR
Data
HIS Feedback iteration mining
Application Evaluation
3
area will not definitely be noise. Furthermore, these pixels are available in the shape of
regular or irregular form. Regular shapes are created because of the sensors defects and
at the same time irregular shape is hopefully the real features or elements of the image.
Apart from that, there are few cases, where the reduction of the impulse is not
considered as a huge problem to collect small detailed information, which are hidden
in a bunch of noise such as astronomical images. Moreover, the imperfect algorithms
induce issues such as leaving some of the noise or removing important features during
processing [10]. The median filtering technique have also comprised of some defects as
well. It is not capable in preserving the edges of the image processed and is relatively
complex and expensive with high means of time. Either the pixels are corrupted or
uncorrupted, it computes each pixel in the same way only. Wide variety of efforts has
been taken by various investigators for reduction and removal of impulse noise in the
medical imaging systems. The algorithms based on the median filter are one of the
benefactions based on the reduction of noise or its removal. These entire algorithms
based on median comprises of their available advantages and complications in pursuing
of images de-noising through requesting later investigation to provide huge
developments in the study area. Median filters are typically applicable even in the day-
to-day industries, and these filters can be utilized in locations, where a small delay can
be authorized in medical image processing. Hence, the complete ignorance of them was
not possible as they are objective and still used. A good denoising algorithm is needed
based on the following factors listed below.
1. Protection of edges: The abnormalities in medical images were effectively
detected by using the medical imaging technique. These complications are mostly
occurred by the mass replication of the cells present inside the human body. To analyze
at what extent the disease could spread, a clear demarcation of cells is needed.
Additionally, the surgeries done by using guidance images, it is essential that the edges
of the region of disease should be recognized accurately. Several post-processing
methodologies will tend to smooth the edges of an image that is not applicable in
medical images. Hence, there is a huge need for the best medical image de-noising
algorithm for securing those edges of the medical images.
2 Preserving the structural similarity: The algorithm for de-noising will not
establish any kind of artifacts, while processing de-noising. It is essential that during
entire minute information are
equivalence with the real image.
4
3. Difficulty: Several applications of medical imaging are considered as the
instantaneous process such as surgeries done under the guidance image information by
utilizing the CT scan images or which needs the instant results present likely in US
imaging. Consequently, it has high significance, where the considered algorithm will
be effective in calculation capacity and comprise of minimum difficulty.
4. Unnecessary of prior databases: The acquiring of medical imaging data is
complex because of its highly privileged nature and related ethical issues. Eventually,
a best de-noising algorithm needs to utilize reduced number of databases to attain a
better de-noising performance. The image comprises of noise, and the image after
removing noise by using de-noising filter is represented in the following Figure 1.2.
5
Input image Noise addition
Restored Filtering
image algorithm
6
despeckling was carried out by using various data elements of image present in
transform domain. The techniques based on multi-resolution or the transform domain
are established for minimizing the points in US image processing.
PCA: The PCA denoising of the MRI relaxometry data is considered as an
enhanced technology for developing the validity of parameters without surrendering
the image resolution. This simple and important step for processing makes the way for
wider applicability of the multi-exponential MRI relaxometry.
WSM: This method was considered as an elaboration of the above-mentioned
analysis of MRI relaxometry. This model proposes a inter-mixing of a 3D sub-bands
wavelets among this filters to enhance the performance of noise removal later, while
maintaining the time for calculation comparable.
Contourlet Transform: The Contourlet transform has been implemented to
conquer the controls of the wavelets transform [12]. It allows several numbers of elastic
directions at every scale, while mostly succeeding the critical sampling.
Wavelet based thresholding techniques: The wavelet thresholding is
considered as simplest non-linear methodology, which processes on single wavelet
coefficient at allotted time. Wavelet thresholding is utilized in Bayesian framework to
process on the statistics of the network signal in non-Gaussian noises. The main
disadvantage in many of the thresholding approaches is the previous understanding of
the number of noise available in the image to calculate an optimum threshold that was
in high demand by them.
1.4.3 Hybrid filters
For connecting the above explained techniques, the latest investigators
significantly aims on the hybrid filters [13]. The flexible filter degrades the noise
present in the images into maximum and minimum frequency elements and it was
minimum amount of speckle content. Contrary method employed in this filtering is the
OBNLM along with equaling of blocks, which is considered as a patch-based recovery
methodology. The advantages of spatial domain filtering and multi-resolution schemes
methodology were utilized by the fast bilateral filter for connecting the wavelet
thresholding. By using the wavelet-based and guided bilateral filter denoising
methodologies, enhanced performance accuracy is achieved.
1.4.4 Low Rank based techniques
The Low-rank techniques are also used as a methodology for denoising. The
7
exploration of the issues related to recovering the despeckled US image using NNM
method was carried out based on inspiration of low rank recovery techniques in
denoising [14]. The difficulties are initially noted by using the WNNM methodology,
which is then altered by including a stage of pre-processing to décor-relate edges of the
ultrasound data in video form by comprising its statistical properties.
1.4.5 Sinogram based methodologies
Sinogram filtering methodologies is utilized to minimize the mixed Poisson
Gaussian noise available in the projected medical data [15]. In this method the process
of filtering is carried out based on projections by utilizing the sonogram or the image
space.
ResNet: This ResNet focuses on the elimination of both artifact and noise by
utilizing the left-over component in-between the LDCT and HDCT images
correspondingly in the training data.
1.5 Medical Storage Systems
The medical information is contained in the form of records accurately and in
detailed manner along with safety assurance, which is considered as an essential part in
medical data management. With the incoming of renewed technologies such medical
records, electronic computing, similar to various application, the domains are highly
depended on computerized storage systems for the process of secured storage and
preservation of health information. Although, in the digital domain, the attackers and
adversaries are normally different than the physical world digitally stored
computerized data can be copied word for word, and records may be revealed to several
types of competitors [16]. To secure the privacy and protection of such computerized
electronic medical records, the privacy and protection laws for consumers are
composed in many countries. This law comprises of rigorous guidelines and needs for
regulations of the medical data storage management. Unluckily, the available storage
features are unable to provide strong safety and privacy assurance mandated by the laws
related to the new digital information domain [17]. For example, various regulations
are needed along with data integrity for the period level of 30 years. But storage of such
records for huge time will need unavoidable change of the hardware and format of the
storage [18]. The resulting transportation to new network servers is loyal and
remarkable. Likewise, if the records require to be replaced after the retention period,
the system of storage must assure safe deletion [19]. Such variety of features is not
available in current storage systems for storing the medical data.
8
The personal medical data is a kind of personal data, which was possessed by
an individual. For protecting the personal private data, the users are allowed to access
the medical record along with proper authorization [20]. The protection of the private
medical records is not only moral liability but also a need from legal guidance. The
traditional computerized medical data storage system comprises of centralized scheme,
which rely on the dependable firewall, such as increased elements of authentication or
acquiring most powerful scheme of encryption and so on [21]. Although, a hacker can
access entire medical data once he/she have got into the health information system that
implies the availability of one point of failure [22]. Meantime, patients have the
mentality to be faithful towards the medical organizations storing medical records. The
securing of medical data of patients has become challenging task to the medical
organizations.
The modern-generation medical data management systems are based on text and
documents. The future-generation medical data storage systems are now under the
process of improvement and it comprises of computerized medical records known as
CPR. The CPR comprises of all sorts of information of medical data and is accessible
online at any time to the care providers with proper authority [23]. The data elements
of the CPR comprise of all types of information objects such images, documents, text,
audio, video between others. The technical complication of this attempt involves the
transmission, visualization, storage, and acquisition of these objects features in an
effective manner [24]. The need for medical data storage management is considered as
a main problem, which is being mentioned by the medical information platforms. The
CPR ensures to provide confidential, accurate, and accessible information to the
relevant healthcare providers. When the patient medical records are stored and retrieved
once digitally, the available and future decision system based on computerized data
systems are used to assist the patients care and liability.
1.6 Benefits of Secured Medical Storage Systems
The health data management is described as the technique of collecting, storing,
retrieving, transferring, and protecting medical data. It allows the healthcare providers
to access the information and utilize it to the workable perceptions [25]. But that is only
half side of it. Let us take a deep study at the various benefits and chances that effective
data storage management can provide to the staffs and patients. The following figure
shows the processing of medical data management system.
9
Medical data Collecting
Storing
Retrieving
Transferring
Protecting
data or consult
Better data analysis: Clear data management and HIE have considered as the
stepping stone for utilizing creative technologies like AI. For example, machine
learning helps in evalua
in particular areas or population groups. It helps the medical care providers to create a
complete view of heath condition of the patients and to carry on the preventive
measures if required.
Improved patient experience
enhanced heath care management systems. The entire data is readily available hence;
they do not need to repeat things again and again. For instance, filling the similar
10
medical forms repeatedly or having the same medical diagnosis repeatedly.
Consolidated storage provides broad entry to their medical care information. The
improved clarity can motivate the activities of patient and increase the performance
accuracy of records. The clear collection of data, storage of data and management in
medical care utilities aims on these three factors to gain advantage from the entire data
collected.
1.7 Multi-disease Prediction Models
Health is considered as main element for the human to survive peacefully. As
the population increases, the disease also increases. Hence, clinicians encountered
difficulties in detecting many numbers of diseases and to accurately identify if the
patients is suffering from a particular disease or not, against the rapid population
growth. Therefore, the huge requirement for the enhanced healthcare systems has been
raised to fulfill the medical requirement of the growing populations. Such technologies
are created with methodologies, data and healthcare that must be preceded in an
accurate, effective, and smart action. Prediction system is considered as the best
procedure that satisfies the expectation level in the required field [26]. There are various
models have been established for detection of single disease. But there exist only few
technologies for the prediction of multi-level disease. The disease identification
technique has aimed to detect multi-level diseases by utilizing the data from medical
records of patients [27]. Apart from the single disease detection problems that only aim
on disease risk management in a model, this study focuses in technically determining
all the available disease of the individual including the chances of disease, which may
develop in future. The figure 1.5 represents the schematic of multi-disease prediction
model.
In day-to-day world, the detection of disease based on the computerized format
by utilizing the clinical data with DL approaches is considered as a broad area of
investigation. But there exist some difficulties in this specific field, which is considered
more remarkable. The multi-level disease prediction based on RNN achieves high
prediction accuracy and it comprises the capability to resolve any size of data, but the
mathematical calculations are difficult to be carried out because of its recurrent nature,
and the learning of RNN is analyzed as complicated one during the operation of large
medical data environments [28]. The multi-task learning is implemented for
and the accurate information results are obtained through
11
this learning process. It also avoids the over fitting problems. Even though, this
proposed method is default to be implemented in huge data experiments. It will succeed
only on the single data experiments.
ensemble learning and deep polynomial networks are found to be helpful even in data,
which are unstructured and this methodology achieves valid predictions but still the
identification and execution of problems is considered as difficult for huge datasets.
When there is a deeper network available, there are issues in optimization. By utilizing
the network analysis and data mining methodologies such as parameter optimization,
tree classification and predictive training-regression, it gains clear accuracy in
predicting diseases and it enhances the information quality. Still, it undergoes some
complications during the processing of real-time datasets and several numbers of
essential resources for the huge management datasets.
Start
Risk factor
identification
Classificatio
Training set
n
Trained Future
Training set prediction
model
Disease No disease
The ML is capable in obtaining calculation time and low robust for over fitting
issues. It maintains more accuracy and it determines the data easily. Although, the
correcting of issues is complicated in particle swan optimization and it comprises of
12
reduced rate for convergence. The GA is considered as computationally expensive,
when compared with the decision trees, and it also has large time consumption. The
FNN will not rely on input data and by reading datasets; the networks are capable to
create the whole network. But, the FNN needs long training portions and there exists
some practical problems [29]. For the determination of CKD, the SVM comprises of
better prediction than ANN and both the methods are utilized in storing the data
information on each cloud network. Eventually, the SVM is considered effective but, it
will not perform huge data applications. Their dependent characteristic for hardware
affects the performance validity. These disadvantages encourage future investigators in
identifying more enhanced techniques for the multi-level disease prediction.
13
categorization of the instance.
K-means clustering: It [30] was a quite abandoned learning that will be utilized
for data information during the presence of unlabeled data. The key element utilized for
the understanding of input sample data is clustering process. The important purpose of
this algorithm is to identify for groups present in data and finding the number of
clusters. For every data point, one of K groups is allotted on the basis of iteration. The
clustering carried out on the points of data is considered based on the identical features.
The output obtained from the K-means algorithm is in the form of clusters and labels
for the data, which are trained.
ANN: ANN is enrolling systems motivated towards the biological neural
networks. ANN comprises of many number of nodes. Each node is referred as
biological nodes present in the brain of humans. These neurons may communicate with
each other by utilizing huge number of links. These considered nodes agree the input
data and they execute straight forward processing on the data. The outcome of this
process is moved to several neurons. The output obtained from each node is known as
node values.
Bayesian Algorithm: The modeling of data mining based on prediction are also
done by utilizing the Bayesian algorithm, it includes a smaller number of mathematical
computations, when compared with the other available algorithms. This algorithm
analyses the probability condition of the input and the output monitored and process
faster analysis indicating the connection between the input and output. This algorithm
is remarkable in its starting stage of detection and utilized in treatment of the harmful
infections for early diagnosis. Figure 1.5 shows the techniques involved in data mining.
Naive Decision
Bayes Tree
Artificial Neural
networks
14
1.9 Deep Learning and Machine Learning in Prediction of Multi-
diseases
The prediction of various diseases is considered as main factor in healthcare
organizations for taking the clear decisions. In health care field, improper affirmations
may leads to delay in the diagnosis or even result in death. Several models for the
purpose of prediction were analyzed and many challenges are affecting the disease
diagnosis. The main challenge considered in healthcare industries is variable
accumulations, where the data resources will be non-simultaneous into remarkable
signs of personal healthcare data. Initially, the healthcare professionals face the
difficulties to collect and estimate huge variety of data for victorious diagnosis and
treatments based on the availability of minimum tools or methodologies. Hence, for the
effective analysis of disease prediction on the basis of the available approaches, only
small number of variables is considered such as gender, height, weight, age and more.
On the contrary, the machine learning utilizes more variables on the basis of calculating
gadgets. Thus, for the disease prediction, the machine learning achieves high
performance in the healthcare environment. The further disease status for future, where
analyzed by utilizing several algorithms. This algorithm paves way to create models for
analyzing data and executing of the obtained results by use of ancient data and the data
from real-time. By implementing machine learning, clinicians choose enhanced
medical record and diagnosis availabilities, which
makes the services of medical industry to improve efficiently. DL is the main
development progress of ML and it is applicable for the collection of main features
from the difficult datasets efficiently from the utilized and ordered and stored learning
approaches. DL provides enhanced performance accuracy in several sectors such as
identification of images, natural language processing, and identification of speech.
LR: LR is an effective and well-known technique for monitored classification
[22]. It is considered as an expansion of normal regression and only the diverging
variable, which generally denotes the appearance or non-appearance of the event. LR
also supports in determining the probability that a renewed occasion included to a
particular class. Since it is a probability, the result takes place in-between 0 and 1.
Hence, for using the LR binary classifier, a threshold is required for allotment to
categorize two classes.
15
SVM: SVM may differentiate both non-linear and linear data. Initially, the
mapping of each data item was carried out with n-dimensional elements space, in which
n is represented as the number of features. Then, they determine the hyperplane that
splits the data into two different sections during the process of maximization of distance
on edges for both the sections and reducing the classification errors. The edge distance
for the sections is considered as the available distance from the decision hyperplane till
its closest specimen, that itself an element of that section. Moreover, every point of the
data was noted first in terms of points in n-dimension space along with the feature value
of each element. To execute the classification, the hyperplane are determined, which
differentiates the two sections by the maximum edge.
CNN: When a neural network comprises of only a fully joined layer, which
arrives a deeper level, the number of parameters will be increased gradually, which will
then maximize the memory footprint and also computational cost is maximized. To
fight against the issues caused by entirely connected layers and improves the
enhancement of neural networks to a deep level, a remedy has identified and
implemented gradually. By utilizing the concept of weight sharing and local correlation,
a CNN is developed, which not only minimizes the amount of parameters, but also
efficiently enhances the training accuracy. Although, CNN was typically established
for computerized vision and it has typically achieved as very powerful tool in many
other fields and the well-known paths including the health field. CNN precedes several
data, where the leading is the operation of images along with the spatial local correlation
data namely health care elements such as CT, chest X-ray and various image data.
Despite, there exists a single data image for the process of disease diagnosis [16], the
CNN model also recognize the medical data processing by utilizing the fusion
technology and feature extraction of various difficult data. Then, the entirely inter-
connected layers are used to succeed in the categorization of disease and it also has high
performance accuracy when comparison of other traditional models.
RNN: The process of task-based dialogue and time series forecasting etc. and
need models to train the features from the arrangements. The complete inter-connection
of neurons is not possible in the fully-connected layers, which can affect the time
management in sequencing. The CNN can resolve the issues related to spatial
dimensions, but the temporal dimension issues will not be resolved effectively. Based
on the above-mentioned complications, the RNN was considered in the network, which
comprises of feedback loops and neurons. For situations such as the prior and next input
16
has dependent in relationship, the RNN indicates specific advantages. The performance
accuracy of RNN is good when correlated machine learning algorithms for the
processing of big data comprised of abnormal noise and delays. But, during the learning
process of RNN, it faces certain issues such as disappearance and explosion of gradient
and also its short-term memory loss, which paves way to introduced improved GRU
and LSTM models.
AE: The availability of huge numbers of unnamed data is originally presented.
It encourages the generation of the unsupervised training algorithms. In real medical
situations, the labeling of some specific names is slight complicated. Hence, some
improved technology is required to categorize or aggregate the medical data to establish
future potential values of the data. Moreover, there are several ways to gain the health
data of patients and it is connected to the real healthcare environment. Some temporary
data loss will occur, and the information will be deformed and incomplete [27]. The
auto-encoder performs effective feature extraction on the unnamed data through the
supervised learning and also operates well in noise removal, feature dimensionality
reduction and restoration of real data [28].
GAN: GAN is a type of model, which is utilized to learn synthetic data on the
basis of the game theory. The concept of this game theory is specifically incorporated
similar to the network learning. Two types of sub-networks are arranged namely
discriminant and generation network. The process utilized by them is similar to the
conflict, and for the continuous optimization of the generation and discriminant
network, the back propagation algorithm were obtained and used. The generation
networks are mainly implemented in training the real distribution samples and
discriminant network is in charge of clearly elaborating the created samples by the
network generation of the true samples. The nature of the creative network is a deep-
CNN and the discrimination network is the CNN.
1.10 Research Motivation
While the specific transformations in the medical environment are provided by
the obtained data, it generates certain challenges. Actually, the inflow of computerized
medical records locates an extra tension on medical caregivers for data management in
an effective manner by assuring the integrity, security and interoperability during the
process of arranging the regulations and policies correspondingly. Few challenges
faced by health clinicians in day- today are discussed below.
17
Fragmented Data: The medical image data is available in terms of structured
or non-structured form and exists in several formats such as digital formats,
images/video/multimedia files, specialized formats, papers or spreadsheets named as
DICOM format, which is used in the process of imaging of MRIs X-rays, and ECGs
MRI scans. The collection of data and aggregation is equally segmented, which makes
the collection and extraction of the data features a real complication. The data are
gathered by the payers, patient, and healthcare providers, but centralizing that
information collected from data has some complications. The duplicate or false data
may evolve, which makes the profiles of healthcare givers to be invalid. The lack of
factors such as supporting, monitoring, and understanding leads to high risks.
Changes to Data and Quality: There will be a constant change in health-care
data. The patients may have several treatments and diagnosis along a particular mode
of time. The diagnosis and operations comprise of overtime, for those new models
known as Telehealth models and new data were generated. The data is the fuel for
operation of medical information system, improving the data integrity means improving
the working of the system potential. The data classified by duplication, invalidity,
inaccuracy, incompleteness and poor standardization is known as dirty data, which may
lead to mistake in data processing and it c -
being, affecting the specialty of healthcare providers and leads to costly lawsuits.
Additionally, the poor-quality data may tend to improper communication internally and
externally, treatment delay and improper payments. The false data may also lead to
affect the potential of proceeding the health research and improvement steps by giving
wrong information.
Data Interoperability: For the proper utilization of the data, it requires
sequential meanings and it needs to transfer among peoples, systems and managements,
this ideology is named as data interoperability. It further, arranges the data of patients
from a range of computerized healthcare systems and allows the patients to take over
the entire control in the sharing of medical records, by improving the research and
creativity. The standardized HER data is needed, which was provided by the medical
care providers but, the medical testing providers working externally uses alternative
systems and protocols. Hence, the sharing of the entire medical record and creating
18
Lack of Integration between Clinical and Administration Systems: There
exist a consolidation gap in-between data of patient and management, internally.
Health-care records obtained by clinicians and on the hospital floor is considered in
insurance claims and billing sections accurately. The configuration of data management
system is carried out to assure that there is a match of diagnosis codes and proper care
provided is accurately monitored for both management purposes and analysis.
Data Security: Data security is considered as the major objective for any kind
of business, specifically the medical care industry. During every movement phase from
one place to another, the data should be resolved. This process makes the data security
a broad-scale concern in day-to-day industries. The data security involves the training
the data security using a potential hardware and software methodologies and the
accuracy and validity of the considered data are assured. The data security concept with
high potential capacity comprises of features such as collection of the needed
information only, protecting the stored data and demolishing any data information that
is not needed further. Due to the increasing cyber-attack cases, invasion of personal
data and data breaches, the management should admit the new and involving privacy
and safety regulations. The breach of sensitive contents leads to criminal and civil
penalty.
Regulations and Compliance: Medical data is a sensitive platform and should
obey the government regulations. The challenges in finding of data and improper
quality of data will make it more complicate to process the audits needed, to meet
regularity needs and control the data diversity of medical care providers, who use the
data governance in medical care organizations helps in better management and usage
of data creates operations for solving the data problems and also allows the users to
make decisions on the basis of highly qualified data. Still, all of this starts with
improved data collection and assuring that the collected data are accurate, updated, and
complete and in reliance with HIPAA regulation standards. A well-implemented
HIPAA issues web form solution can be indispensable in allowing the healthcare
managements to manage data collection, along with new forms for patients, HIPAA
release forms, patient medical history forms, consent forms and contact update forms.
The process of digital imaging plays a significant process in analysis of medical
data, which supports in identifying information that is free from mistakes and provides
precise identification and diagnosis. Various image processing methodologies have
19
been used in the health care environment for data analysis. The selection of proper
methodology to perform effective de-noising without any information loss or changing
of edges is considered as a complicate process. The aim lies behind the attainment of
noise free medical images is to preserve the edges of tissues and boundaries at same
time in securing small changes in the same tissue areas. There is no method obtained to
be effective till now considering all factors such as edge preservation, robustness, noise
reduction, elimination, cost of computation and user interaction. The algorithms such
as adaptive filters create promising results for images of brain disease but gives bad
results for other disease images known as cancer images and X-Rays. Hence, the
process of noise removal from these data images is an essential and major step in the
digital image processing technique. The type of noise present in the image analyzes the
model in denoising. The method of denoising mainly aims in enhancing the image
quality by improving indication to noise ratio and securing the minute details
specifically. For eliminating the degradation effects takes place in the captured image,
several image restoring methodologies and techniques are obtained, which is used in
minimizing, eliminating or removing the noise present in data image. The image quality
tends to decompose due to low clarity and due to noise available or by blurring effects.
Blurring takes place because of the imperfect, incomplete or incorrect formation of
images, where the noise minimizes the visual quality of data image by disturbing the
original signals.
20
processing, an adaptive noise removal framework is implemented for enhancing the
performance of the model. The comparative analysis is processed between SNR and
peak SNR and as a result, it achieved high accuracy in noise removal and also
ensures the future scope of the project.
To develop a new medical image storage system with improved security and
privacy protection during the process of transferring the medical images by utilizing
the optimized deep learning architecture along with the support of an enhanced
optimization algorithm and to implement a hybrid chaotic map named HLPWCM
by combining the 2DLCM and PWLCM for image encryption with a deep learning-
based key generation for securing the medical images. Then, enhanced optimization
algorithm named BF-CSHO is introduced for optimizing the weights of RNN-based
key generation for improving the overall performance by minimizing the difficulty
in image transmission. Thus, the experimental assured effective performance in
noise reduction.
To develop the process of multi-disease prediction by utilizing the improved DL
concept. For this, several datasets in relation with different diseases are collected
and pre-processing was carried out with data normalization avoiding the redundant
data. Then, the proposed JA-MVO algorithm is implemented by combined meta-
heuristic algorithms such as MVO and JA and the weighted feature extracted and
the weights of DBN and RNN were optimized. Hence, the resultant analysis
provides high performance in developed method, when comparing the available
traditional models for multi-disease prediction.
21
research gaps and challenges related to the medical data storage system were
provided.
Chapter 3 describes the improved de-noising techniques for medical images by
collecting images and analyzing them by using proposed models and then, the
results were compared for accuracy and further recommendations were provided.
Chapter 4 explains the Optimized Deep Learning model in medical data storage and
processing and the HLP-WCM were done to secure the stored medical data. The
experimental results provide high performance in effective noise removal of
medical data images.
Chapter 5 implements an improvised DL concept for multi-disease prediction
model. Additionally, the JA-MVO algorithms were proposed to provide better
results for the noise removal of medical data storage system. The results provide
superior performance in proposed model.
Chapter 6 concludes the medical data processing model along with its future scopes
and objectives.
22