Applications of Machine Learning in Healthcare
Applications of Machine Learning in Healthcare
7,200
Open access books available
190,000
International authors and editors
205M Downloads
154
Countries delivered to
TOP 1%
most cited scientists
12.2%
Contributors from top 500 universities
Abstract
1. Introduction
1
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Artificial intelligence (AI) has been intricately linked to the rise of modern-day
computing machines. Machine learning has its roots and beginnings firmly planted
in history. Alan Turing’s work in cracking the German Enigma machine during
World War II became the basis for much of modern computer science. The Turing
Test, which aims to see if AI has become indistinguishable from human intelligence,
is also named after him [1, 2].
At the height of the Second World War, the Allies had a significant logistical
hurdle in the Atlantic. The United States and United Kingdom needed to set up
secure shipping lines to move both armaments and troops to England in preparation
for a mainland European invasion. However, the German U-boats were extremely
effective at disrupting and sinking many of the ships traversing these shipping lanes
[3]. As such, the Allies needed to intercept German communications to swing the
2
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
Battle of the Atlantic in their favor. The Germans encrypted their communications
with The Enigma Machine, the most sophisticated encryption device of its time.
Turing and the rest of Bletchley Park were tasked with breaking the coded
messages produced by The Enigma Machine and eventually produced The Bombe,
a mechanical computing device which successfully decoded the cipher of The
Enigma machine (Figure 1). Using the Bombe, they read the German orders sent to
submarines and navigated their ships around these dangers. This was Turing’s first
intelligent machine. Alan Turing would later go on to describe the idea of a thinking
machine which would eventually be called AI [4].
Machine learning is a subset of AI and the term was coined in the late 1950s by
Arthur Samuel who published a paper on training computers to play checkers when
he worked with IBM [5]. AI is best described as giving human-like intelligence to
machines in a manner that directly mimics the decision making and processing of
the human conscience. ML is the subset of AI that focuses on giving machines the
ability to learn in an unaided manner without any human intervention.
By the late 1960s, researchers were already trying to teach computers to play
basic games such as tic-tac-toe [6]. Eventually, the idea of neural networks,
which were based on a theoretical model of human neuron connection and com-
munication, was expanded into artificial neural networks (ANNs) [7, 8]. These
foundational works laid dormant for many years due to the impracticality and poor
performance of the systems created. Computing technology had not yet advanced
enough to reduce the computational time to a practical level.
The modern computer era led to exponential increases in both computational
power and data storage capacity. With the introduction of IBM’s Deep Blue and
Google’s AlphaGo in recent decades, several leaps in AI have shown the capacity of
Figure 1.
Picture of the German Enigma machine which was used to code military communications. Taken from
Wikimedia Commons.
3
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
AI to solve real world, complex problems [9, 10]. As such, the promise of machine
learning has taken hold in almost every sector imaginable.
The widespread adoption of machine learning can be mostly attributed to the
availability of extremely large datasets and the improvement of computational
techniques, which reduce overfitting and improve the generalization of trained
models. These two factors have been the driving force to the rapid popularization
and adoption of machine learning in almost every field today. This coupled with the
increasing prevalence of interconnected devices or the Internet of Things (IoT) has
created a rich infrastructure upon which to build predictive and automated systems.
Machine learning is a primary method of understanding the massive influx of
health data today. An infrastructure of systems to complement the increasing IoT
infrastructure will undoubtedly rely heavily on these techniques. Many use cases
have already show enormous promise. How do these techniques work and how do
they give us insight into seemingly unconnected information?
4
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
Figure 2.
Example of a workflow for training and assessing a random forest model. Each green triangle represents an
independently trained tree from the training data. The prediction of each tree is summed and is represented
as the model. Test data is then fed to the model, i.e., all the trees, and the resulting prediction is made. The
prediction is then compared to the original test data to assess how the model performs.
Typically, better results can be achieved with gradient boosting, but tuning is much
more difficult, and the risk of overfitting is higher. Gradient boosting works well
with unbalanced data and training time is significantly faster due to the gradient
descent nature of the algorithm [17, 18].
5
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Figure 3.
Example of a simple workflow for training and assessing a gradient boosting machine model. Each green
triangle represents a trained tree from the training data with the subsequent tree using the residuals or errors
from the prior tree to improve its prediction. The prediction of each tree is summed and is represented as the
model. Test data is then fed to the model, i.e., all the trees, and the resulting prediction is made. The prediction
is then compared to the original test data to assess how the model performs.
immediately obvious. However, they are also more computationally intensive and
require a larger amount of data to perform.
The most common and well-known algorithms are K-means clustering and deep
learning, though deep learning can be used in a supervised manner [12, 20]. Such
algorithms also perform association tasks which are similar to clustering. These
algorithms are considered unsupervised because there is no human input as to what
set of attributes the clusters will be centered on.
The typical k-means algorithm has several variations such as k-medians and
k-medoids, however the principle is the same for each algorithm. The algorithm
uses Euclidian distance to find the “nearest” center or mean for a cluster assuming
there are k clusters. It then assigns the current data point to that cluster and then
6
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
recalculates the center for the cluster, updating it for the next data point [21]. The
biggest drawback to this algorithm is that it must be initialized with an expected
number of “means” or “centers.” Improper selection of the k value can result in poor
clustering.
Deep learning uses neural nets to perform predictions even on unlabeled data as
well as classification techniques. Based off models of human neurons, perceptrons,
as they are typically called, are organized into many networked layers making the
network “deep” in nature [20]. Each perceptron has multiple inputs and a single
output. They are organized into layers where the outputs of the previous layer
serve as the inputs for the next layer. The input layer requires one perceptron
per input variable and the subsequent layers are determined before training by
a human (Figure 4). This is one of the difficulties and challenges in building an
effective neural net. The computationally intensive nature of computing each
perceptron for a large neural net can mean that training alone can take days to
weeks for large data sets [22].
2.1.3 Hyperparameters
Figure 4.
Example of a simple neural net with two hidden layers of three perceptrons each. The number of inputs,
number of hidden layers, and number of perceptrons in each layer can be changed. Additionally, the
connections between layers and perceptrons can also be changed.
7
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
be the maximum number of leaves in a tree or the maximum depth for a tree. Other
common hyperparameters include learning rate, batch size, dropout criterion, and
stopping metric.
Properly selecting hyperparameters can significantly speed up the search for
a proper generalized model without sacrificing performance. However, in many
cases finding the proper set is more of an art than a science. Many researchers have
attempted to make hyperparameter searching a more efficient and reproducible
task [23–25]. Again, this process also highly depends on the algorithm, dataset, and
problem you are trying to solve. A machine learning model can be tuned a nearly
infinite amount of different ways to achieve better performance. Hyperparameters
represent a way to reproduce results and also serve as a tool to properly vali-
date models.
Considering the pace of research in the field, there are constant advances
and improvements to many of these machine learning techniques, but the
important thing to remember is that not all algorithms work for all use cases.
Each algorithm has advantages and disadvantages. Certain data types may also
affect the performance of individual algorithms and the time spent implement-
ing such models will often be a result of testing different variations, parameters,
and hyperparameters within these algorithms to achieve the best generalized
performance.
The goal of any machine learning algorithm is to utilize real data to create a
model that performs the best on real-world scenarios, and that can be assessed in
a quantitative, reproducible manner. Assessment of statistical models is a whole
subfield in itself, but we will briefly discuss the basics, which are applicable for
almost any machine learning algorithm you will come across.
This is also sometimes referred to as the recall or hit rate, or just simply the true
positive rate, and the sensitivity is equivalent to 1 − False Negative Rate.
Specificity is the probability of a negative result given that the sample is nega-
tive. Mathematically,
Number of True Negatives
Specificity = _______________________________________
Number of True Negatives + Number of False Positives
This value is also referred to as the selectivity of the test. This is equivalent to
1 − False Positive Rate.
8
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
2.2.3 Overfitting
Overfitting is one of the main concerns when training any model [26]. Simply
put, when training a model on a set of data, over-training the model will improve
the performance of the model on that specific dataset but at the cost of losing
generalization to other datasets. An overfitted model will not work when applied to
new data it has never seen before. From a practical standpoint, such a model is not
very useful in a real-world application.
When training any machine learning model, the ideal result is a generalized
model. A generalized model works well on a variety of different cases and a vari-
ety of different datasets, especially data it has never seen before. As such, many
researchers are hesitant to give too much credence to a model or method that utilizes
a single dataset.
A variety of methods have been used to prevent models from overfitting and
many of these are now encapsulated in the hyperparameters discussed earlier.
9
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Figure 5.
Examples of an AUC denoting a model which has good predictive power (left) and an AUC denoting a model
with poor or near random predictive power (right). 1 – Specificity is sometimes written as false positive rate
( fpr) and sensitivity can be read as true positive rate (tpr).
The idea is to prevent the models from adapting too quickly to the dataset it is being
trained on. This subset of methods is known as regularization [27].
One such method, used in neural nets, is called dropout. This method is widely
used to prevent artificial neural nets from overfitting during classification tasks.
The method is fairly simple. During the training process, random perceptrons and
their corresponding connections are “dropped” from the network. These “thinned”
networks have better performance compared to other regularization techniques on
supervised learning tasks [28].
Often a method known as cross-validation is used to assess the performance and
validate the generalized predictive ability of a model. The most common method for
building machine learning models is the partitioning of the data set into roughly
80% for training and 20% for testing. This partition is typically less useful for linear
models but splitting is more beneficial for complex models [29]. During cross-val-
idation, this split is done in separate sections of the data to ensure proper coverage.
For example, if a 10-fold cross-validation is performed, the first split in a data set
with 100 observations could be a the first 80 for training and the last 20 for test, the
second split could be the first 10 and last 10 for test and the middle 80 for training,
etc. (Figure 6). This creates 10 models using the same algorithm just trained and
tested on different portions of the same data. The average performance of these 10
models gives a good measurement of the generalized performance of the algorithm
on that type of data.
The healthcare sector has always had a very large amount of information, often
times stored as physical documents in clinics, hospitals, regulatory agencies, and
biomedical companies [30, 31]. With the push to electronic medical records (EMR),
this information is rapidly being transformed into a form which can be leveraged
by AI technologies. The estimated amount of healthcare data stored in 2011 was
around 150 exabytes (1 EB = 1018 bytes), though that number is most likely expo-
nentially larger almost a decade later [32, 33]. These large databases, when in a
digitized form, are often known as Big Data.
However, such healthcare information is very different in both form and func-
tion. Visual data in the form of medical images is very different than familial history
10
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
Figure 6.
Example of a set of cross validation splits. There are n splits for the number of iterations desired and the results
of all iterations are averaged to assess the generalized performance of a model trained on a dataset.
which may be simple text-based information. Laboratory and clinical tests may be
reported as numbers only, while health outcomes are often qualitative in nature
and may be a simple yes or no entry in a spreadsheet. Insurance and administrative
data is also indirectly linked to various information, such as patient outcomes, while
information from sensor based technologies like EKGs, pulse oximeters, and EEG
provide time-series data of vital signs [34].
Additionally, the genomic revolution has contributed enormously to the data
explosion. Large-scale genetic databases such as the Cancer Genome Atlas (TCGA)
and the UK Biobank include thousands of patients’ genetic sequencing information
along with various other health information such as disease state, age of diagnosis,
time of death, and much more [35–38]. Copy number variation (CNV) data from
the UK Biobank’s roughly 500,000 patients, which does not even contain the raw
sequence reads, is almost 2 Terabytes (TB) alone in flat text files. These genetic
databases rely on an array of assays and sequencers spread across different hospitals
Table 1.
Overview of largest biobank databases as of 2019.
11
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
and research facilities around the globe, before being processed and transferred to
their respective centralized storage databases [35, 39, 40].
The collection of biological data and creation of these databases show no
evidence of slowing down. Many biobanks, databases which contain some form
of biological samples such as blood or serum, contain thousands of participants
and many have plans to collect hundreds of thousands of samples from patients
(Table 1). Because many databases are growing so quickly it is unclear how much
data resides in many of these databases. However, The Cancer Genome Atlas alone
contains 2.5 petabytes (1 PB = 1015 bytes) of data and the UK Biobank contains 26
terabytes (1 TB = 1012 bytes) of just genetic information (UK Biobank also contains
medical images such as brain scans which is not included in this table).
Implementing machine learning systems into a hospital with this complex
information Web is usually slow, due to the abundance of caution needed to ensure
patient health. Many physicians are also wary of adopting new systems that are
unproven in a clinical setting due to the risk of litigation and potentially cata-
strophic consequences for their patients.
12
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
the ability to make such databases open-source. Many studies only use around
100–1000 samples in training CNNs. This limited sample size increases the risk of
overfitting and reduces the accuracy of the predictions [56].
Concerns regarding the implementation of machine learning into clinical diagno-
sis have been raised regarding proper validation of models [57]. The main fears entail
properly scoping the intended goals of a machine learning model, reducing dimen-
sionality of the data, and reproducibility of training such models on real-world and
new clinical data. Validating results on other datasets can be difficult due to the lack
of larger datasets for niche diseases, where the aggregation of this data can take more
work than the actual training of the model. Medical imaging data is inherently more
difficult to acquire and is more difficult to store and process. The infrastructure to
handle the data has simply not kept up with the increase in the amount of data.
Figure 7.
Example of mammogram with the left image being that of a raw mammogram and the right hand being the
image with the detection overlaid with the region of interest in white, using NASA software originally used to
enhance earth science imagery. Taken from NASA press release, credited to Bartron Medical Imaging.
13
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Electronic medical records (EMR), the new standard in many hospitals, require
complex digital infrastructure. Unification of health data in a formatted manner is a
major goal as it should increase the efficiency of hospitals as well as improve patient
health outcomes. However, a significant problem is the historical existing physical
documentation. Transferring these existing documents into an electronic form is
difficult and would be very tedious and expensive if people were hired to manually
input such information into an electronic system.
One application of machine learning, which may aid in this problem, is natural
language processing (NLP). By scanning these documents rapidly and integrating
the resulting images into a database, these systems attempt to extract readable data
from free text and incorporates image processing to identify key words and terms.
Handwritten physician notes contain information such as patient complaints, the
physicians own observations, and patient family history. This clinical information
can be annotated. However, poorly worded or inaccurate writing by the physician
can make it difficult to accurately assign this information to appropriate categories.
Forms and documents that already have structure make for much easier language
processing, though there is still the risk of missing data Figure 8.
Creating a system for improved clinical decision support (CDS) with old patient
records is feasible. Any such system is structured to aid in clinical decision making
for individual patients based on a database of computerized knowledge. Such a
system could be envisioned as two-fold: 1. extracting facts about the patient from
their medical record, either through written or typed physician notes or labs or dic-
tation involving audio NLP, 2. Associating possible disease states based on extracted
information from previous known cases or through literature search via NLP [62].
Integration of several specialized NLP systems is required for any true and practical
implementation of such a CDS system.
Likewise, compilation of the existing scientific research into central repositories
is a difficult task. Sometimes physicians may be unaware of a promising new treat-
ment just due to the difficulty of parsing the tidal wave of new papers. Scientific
publications have always been widely dispersed across multiple journals and the
modern-day information explosion has only exacerbated the issue. When it comes
to compiling information such as results from genome-wide association studies
(GWAS), the primary method has been a manual curation of the information by
certain individuals within the scientific community: “librarians” so to speak.
Recently, a paper published in Nature Communications used machine learning
systems to automatically compile GWAS information from open-access publications
14
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
Figure 8.
Example of a nursing care plan which represents a formatted health document. Most of these plans were filled
out by hand and many hospitals have transitioned such forms to electronic records. However, older documents
still need to be transferred to digital form. Taken from Wikipedia commons.
and extract GWAS associations into a database with the aim of helping curators.
Though the results are somewhat inconsistent (60–80% recall and 78–94% preci-
sion) it represents one of the many ways NLP is being utilized to aid in medical
discovery [63].
There are many exciting possibilities where NLP could be used to improve medi-
cine and medical research. We will discuss a few interesting findings with similar
approaches but different goals. This is by no means an expansive list but highlights
the broad spectrum of possible machine learning applications.
In 2015, a research group published a paper reporting 100% accuracy of predict-
ing onset of psychosis using recorded dialog of clinically high-risk youth. Each
youth was interviewed over a period of 2.5 years every 3 months. Based on the
15
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Genetic information and technologies have exploded since 2008, creating dif-
ficult challenges in how to handle the exponentially increasing data. Advances in
genetic sequencing speed, namely NGS technologies have exponentially increased
the speed at which a whole human genome is sequenced, while also dramatically
reducing costs. The human genome is a complex physical structure that encodes all
the information of human development and characteristics. The genome is highly
interconnected and deciphering most of these instructions is still a mystery to us.
Variation of genomes between people also increases the complexity of understand-
ing gene interactions.
Many health initiatives have focused on acquiring large sample sizes of human
genomes to help identify statistically relevant trends among different populations
of humans. However, the 23 chromosomes of the human genome contain around
20,000 genes which have been identified as the primary coding sequences for the
proteins necessary in building the biological components of our cells [67]. This
number is still a rough estimate and some estimates indicate that there may be as
many as 25,000 genes or as few as 19,000 [68, 69]. A large swathe of genetic infor-
mation that does not code for any proteins is not included in these estimates.
A growing body of literature indicates that certain sections of what has been
colloquially called genetic dark matter, or missing heritability, exists [70–74]. These
terms refer to the portions of DNA which have no apparent protein coding function,
but may be relevant to the level of gene expression in a person’s genetic code [75, 76].
Levels of gene expression may cause protein overload or deficiency, which can lead
to a variety of health problems. Additionally, structural differences in the physi-
cal structure of how the DNA is bound into chromosomes and then subsequently
unwrapped during both the duplication process and translation and transcription
process, can also affect the level of gene expression.
16
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
For example, methylation or acetylation of the DNA backbone can make it more
difficult (methylation) or easier (acetylation) to unravel the DNA strand during
normal cell processes like replication or protein assembly. Evidence of multiple
copies of the same gene have also been classified in what is described as copy
number variations (CNV) which indicate duplication, triplication, and deletion
events of certain areas of the genome in an individual. Understanding this highly
interconnected and nonlinear relationship between all the different of the areas of
the human genome is difficult.
With machine learning, scientists have begun to find patterns and trends which
can be modeled in a more predictable manner. Utilizing the ever-growing amount
of genetic data, machine learning has the potential of accurately predicting who is
at risk of acquiring certain diseases such as cancers and Alzheimer’s disease. Mental
illnesses such as schizophrenia and bipolar disorder have also been known to run in
families, indicating a possible genetic link.
Disease risk can be broadly categorized into inherited risk and environmental
risk. Inherited risk describes a person’s disposition to acquiring complex diseases
due to a trait which is genetically passed down from their predecessors. This
includes genetic mutations contained within their germline DNA which may predis-
pose them to cancers or other health conditions [77, 78].
Environmental risk describes somatic mutations, or mutations to a person’s DNA
due to something they have encountered in their environment. These mutations can
still increase a person’s risk of acquiring a disease but they do not affect the germ-
line, and will not be passed on to their progeny and thus will not be inherited [79].
Inherited risk describes mutations that exist in the human germline and which
will be passed onto the offspring through normal reproduction. Whereas, somatic
mutations may affect organs or a set of cells, germline mutations exist in all the cells
of the offspring. Many of these mutations may be passed through paternal lineage
and there is some indication that certain individuals may have disease predisposi-
tion but which cannot be directly linked to familial history but could still be due to
these hidden germline mutations [80–82].
Several different types of mutations may exist within a human genome. They are
broadly categorized as single nucleotide polymorphisms (SNPs), structural varia-
tions or copy-number variations (CNVs), and epigenetic variations.
SNPs are a single or point mutation of one base pair in the human genome that
occurs in at least 1% of the human population [83, 84]. These mutations are the
most common source of genetic variation and can occur both within coding regions
and outside of coding regions of the genome. SNPs contribute to vast differences
even between relatives and can arise because of both inheritance and development
in the womb. Within SNPs there are common and rare variants, with rare variants
occurring less than 0.5% within the global sample [84].
Structural variations and specifically CNVs are deletions, insertions, duplica-
tions, and inversions of large regions of DNA. These structural differences are
usually inherited and a typical human can have anywhere between 2100 and 2500
structural variations [84]. These variations were found to cover more of the human
genome than SNPs alone [83].
Epigenetic variation describes variations in the chemical tags attached to
DNA or associated structures such as histones, which affects how genes are read
and activated. Epigenetics includes DNA methylation and acetylation, histone
modifications, and non-coding RNAs which all affect the degree to which a gene
may be expressed [85]. As a newer field, it is unclear how much of these epigenetic
17
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Table 2.
Sampling of performance of GBM models trained on data from the Cancer Genome Atlas.
variations are inherited from generation to generation, and how much is a result of
environmental factors [86].
6. Conclusions
18
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
Conflict of interest
Notes/thanks/other declarations
The author would like to thank the University of California, Irvine for support
during the writing of this chapter.
19
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
Author details
© 2020 The Author(s). Licensee IntechOpen. Distributed under the terms of the Creative
Commons Attribution - NonCommercial 4.0 License (https://creativecommons.org/
licenses/by-nc/4.0/), which permits use, distribution and reproduction for
non-commercial purposes, provided the original is properly cited.
20
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
References
[1] Copeland J. The turing test. Minds Cliffs, NJ. The Knowledge Engineering
and Machines. 2000;10(4):519-539 Review. 1996;11(1): 78-79
[2] French RM. The turing test: The first [12] Alpaydin E. Introduction to Machine
50 years. Trends in Cognitive Sciences. Learning. London: The MIT Press.
2000;4(3):115-122 2014;3:640
[3] Edwards S. World war II at sea: A [13] Kotsiantis SB. Supervised machine
global history. The Journal of American learning: A review of classification
History. 2019;106(1):237 techniques. Informatica (Ljubljana).
2007;31(3):249-268
[4] Turing AM. Computing machinery
and intelligence. Parsing the Turing [14] Hastie T, Tibshirani R, Friedman J.
Test: Philosophical and Methodological The Elements of Statistical Learning -
Issues in the Quest for the Thinking Data Mining, Inference, and
Computer. 2009. p. 23-65 Prediction. Springer Series in Statistics.
Switzerland: Springer. 2009;2(1):93-85
[5] Samuel AL. Some studies in machine
learning. IBM Journal of Research and [15] Ho TK. A data complexity analysis
Development. 1959;3(3):210-229 of comparative advantages of decision
forest constructors. Pattern Analysis &
[6] Samuel AL. Programming computers Applications. 2002;5(2):105-112
to play games. Advances in Computers.
1960;1(C):165-192 [16] Friedman JH. Stochastic gradient
boosting. Computational Statistics and
[7] Fukushima K. Cognitron: A Data Analysis. 2002;38(4):367-378
self-organizing multilayered neural
network. Biological Cybernetics. [17] Friedman JH. Greedy function
1975;20(3-4):121-136 approximation: A gradient boosting
machine. The Annals of Statistics.
[8] Fukushima K. Neocognitron: 2001;29(5):1189-1232
A self-organizing neural network
model for a mechanism of pattern [18] Mason L, Baxter J, Bartlett P,
recognition unaffected by shift in Frean M. Boosting algorithms as
position. Biological Cybernetics. gradient descent in function space. NIPS
1980;36(4):193-202 Conference Proceedings. 1999:512-518
[9] Huang K, Mashl RJ, Wu Y, Ritter DI, [19] Libbrecht MW, Noble WS. Machine
Wang J, Oh C, et al. Pathogenic germline learning applications in genetics and
variants in 10,389 adult cancers. Cell. genomics. Nature Reviews. Genetics.
2018 2015;16(6):321-332
21
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
[26] Hawkins DM. The problem of [35] The Cancer Genome Atlas
Overfitting. Journal of Chemical Program—National Cancer Institute
Information and Computer Sciences. [Internet]. 2019. Available from:
2004;44(1):1-12 https://www.cancer.gov/about-
nci/organization/ccg/research/
[27] Ng A. Regularization—Solving the structural-genomics/tcga
Problem of Overfitting. Coursera; 2011.
Available from: https://www.coursera. [36] tcga-infographic-enlarge.__
org/learn/machine-learning/lecture/ v100169753.png (1400×2580)
ACpTQ/the-problem-of-overfitting [Internet]. 2019. Available
from: https://www.cancer.gov/
[28] Srivastava N, Hinton G, PublishedContent/Images/images/nci/
Krizhevsky A, Sutskever I, organization/tcga/tcga-infographic-
Salakhutdinov R. Dropout: A simple enlarge.__v100169753.png
way to prevent neural networks from
overfitting. Journal of Machine Learning [37] Peakman TC, Elliott P. The UK
Research. 2014;15(56):1929-1958 biobank sample handling and storage
validation studies. International Journal
[29] Picard RR, Cook RD. Cross- of Epidemiology. 2008;37(2):234-244
validation of regression models. Journal
of the American Statistical Association. [38] Sudlow C, Gallacher J, Allen N,
1984;79(387):575-583 Beral V, Burton P, Danesh J, et al. UK
biobank: An open access resource for
[30] Jothi N, Rashid NA, Husain W. identifying the causes of a wide range of
Data mining in healthcare—A review. complex diseases of middle and old age.
Procedia Computer Science. PLoS Medicine. 2015;12(3)
2015;72(1):306-313
[39] Protocol for the UK Biobank—
[31] Koh HC, Tan G. Data mining Wayback Machine [Internet]. 2019.
applications in healthcare. Journal of Available from: https://web.archive.
22
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
[40] Elliott P, Peakman TC. The UK [49] Scholtens S, Smidt N, Swertz MA,
biobank sample handling and storage Bakker SJL, Dotinga A, Vonk JM, et al.
protocol for the collection, processing Cohort profile: LifeLines, a three-
and archiving of human blood generation cohort study and biobank.
and urine. International Journal of International Journal of Epidemiology.
Epidemiology. 2008;37(2):234-244 2015;44(4):1172-1180
[41] Gao GF, Parker JS, Reynolds SM, [50] The Health 202: NIH wants 1
Silva TC, Wang LB, Zhou W, et al. million Americans to contribute
Before and after: Comparison of legacy to new pool of gene data. The
and harmonized TCGA genomic Washington Post [Internet]. 2019.
data commons data. Cell Systems. Available from: https://www.
2019;9(1):24-34.e10 washingtonpost.com/news/powerpost/
paloma/the-health-202/2018/01/16/
[42] Bycroft C, Freeman C, the-health-202-nih-wants-1-
Petkova D, Band G, Elliott LT, Sharp K, million-americans-to-contribute-to-
et al. The UK biobank resource with new-pool-of-gene-data/5a5ba45a30fb04
deep phenotyping and genomic data. 69e8840135/
Nature. 2018;562(7726):203-209
[51] FACT SHEET: President Obama’s
[43] Background—EPIC [Internet]. Precision Medicine Initiative.
2019. Available from: http://epic.iarc.fr/ whitehouse.gov [Internet].
about/background.php 2019. Available from: https://
obamawhitehouse.archives.gov/
[44] Leitsalu L, Haller T, Esko T, the-press-office/2015/01/30/fact-
Tammesoo ML, Alavere H, Snieder H, sheet-president-obama-s-precision-
et al. Cohort profile: Estonian biobank medicine-initiative
of the Estonian genome center,
university of Tartu. International [52] Precision Medicine Initiative (PMI)
Journal of Epidemiology. Working Group. The precision medicine
2015;44(4):1137-1147 initiative cohort program—Building
a research foundation for 21st century
[45] Master Decoder: A Profile of Kári medicine. Precision Medicine Initiative
Stefánsson | The Scientist Magazine® Work Group Report to Advisory
[Internet]. 2019. Available from: Committee to Director NIH; 2015
https://www.the-scientist.com/profile/
master-decoder--a-profile-of-kri- [53] FinnGen, a global research project
stefnsson-65517 focusing on genome data of 500,000
Finns, launched. EurekAlert! Science
[46] Gulcher J, Stefansson K. Population News [Internet]. 2019. Available from:
genomics: Laying the groundwork for https://www.eurekalert.org/pub_
genetic disease modeling and targeting. releases/2017-12/uoh-fag121917.php
Clinical Chemistry and Laboratory
Medicine. 1998;36(8):523-527 [54] Le QV. Building high-level features
using large scale unsupervised learning.
[47] China Kadoorie Biobank [Internet]. In: ICASSP, IEEE International
2019. Available from: https://www. Conference on Acoustics, Speech and
ckbiobank.org/site/ Signal Processing—Proceedings. 2013
[48] Littlejohns TJ, Sudlow C, Allen NE, [55] Greenspan H, Van Ginneken B,
Collins R. UK biobank: Opportunities Summers RM. Guest editorial deep
23
Smart Manufacturing - When Artificial Intelligence Meets the Internet of Things
24
Applications of Machine Learning in Healthcare
DOI: http://dx.doi.org/10.5772/intechopen.92297
25