Deep Learning Techniques For Electronic Health Record Analysis
Deep Learning Techniques For Electronic Health Record Analysis
Deep Learning Techniques For Electronic Health Record Analysis
net/publication/330876062
CITATIONS READS
2 94
4 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Eugenio Vocaturo on 28 June 2019.
Abstract—Following the general trend, the amount of digital reporting these is out of the scope of this paper. Bottom line,
information in the stored electronic health records (EHRs) had an the following seven V’s provide a good definition of Big Data.
explosion in the last decade. EHRs are not anymore used, as in the
past, to store basic information of the patient and administrative • Volume. Volume is the total amount of data. In the recent
tasks, but they may include a range of data, including the medical past it was measured in Gigabytes, now Zettabytes (ZB)
history of the patient, laboratory test results, demographics, or even Yottabytes (YB) are used.
medication and allergies, immunization status, radiology images,
vital signs. At the present, the problem has shifted from collecting
• Velocity. Velocity is the speed in which data is produced,
massive amounts of data to understanding it, i.e. use EHRs processed and analyzed. Making quick decision is crucial
for turning data into knowledge, conclusions and actions. EHRs for a company and avoids stasis.
were not designed to forecast disease risk or disease progression • Variety. Variety describes the different format of data that
or to determine the right treatment, but if they are combined cannot be stored in structured relational database systems.
with artificial intelligence (AI) algorithm this issue became
possible. The need for tools allowing to construct predictive
Organizing rapidly changing data having different format
models capturing disease progression is a priority. In the recent is one of the biggest challenge of Big Data.
past EHRs were analyzed using traditional machine learning • Variability. Variability is different from variety. It has
techniques, whereas recently the progress in the field of deep several potential meaning. Is the data consistent in terms
learning let to the application of deep learning techniques to of availability at times? In the presence of extreme values
EHRs. This paper reports a brief overview of some recently
developed deep learning tools for EHRs.
are these outliers or just noisy data?
Index Terms—Electronic Heath Record, Deep Learning, Pre- • Veracity. Veracity refers to the accuracy of the data.
dictive Tools. Data have to conform with truth, therefore key questions
are related to the provenance of data, to the reliability,
I. I NTRODUCTION accuracy and completeness of sources.
recent study of the IBM [1] states that about the 90%
A of the data in the world has been created in the last two
years [1]. The current output of data is roughly 2.5 quintillion
• Visualization. Visualization refers to the need of tools
to figure out data and analysis’ results on data. Having a
vast amount of data is useless if you are not effective in
bytes a day. The IDC white paper by Seagate [2] forecasts that capturing the meaningful part of it.
data creation will swell to a total of 163 zettabytes (ZB) by • Value. Value is the final aim of Big Data. Each organi-
2025. Seagate’s report also predicts that in the next future the zation has to get value from the data.
primary creators of the world’s data will not be consumers, but
Following the general trend, the amount of digital informa-
enterprises that will create the 60% of the world’s data in 2025.
tion in the stored EHRs had an explosion in the last decade.
This indicates that the very next will be an era focused on the
EHRs are not anymore used, as in the past, to store basic
value of data: creating, utilizing and managing life critical data
information of the patient and administrative tasks, but they
necessary for the smooth running of daily life for consumers,
may include a range of data, including the medical history of
governments and businesses. Big Data are everywhere! They
the patient, laboratory test results, demographics, medication
are part of the information processing system of all sectors,
and allergies, immunization status, radiology images, vital
from science to government, from health care to media, from
signs.
university to real time commerce, from security/privacy to
financial [3], [4], [5]. But what is big data? Many different At the present, the problem has shifted from collecting
definitions have been provided in the recent literature and massive amounts of data to understanding it, i.e., turning data
into knowledge, conclusions, and actions. One of the major
task of modern medicine is the possibility of anticipating the
978-1-5386-8161-9/18/$31.00
2018
c IEEE future evolution of a disease, by predicting the risks for the
patient so that helping in prevention. This specific task is often Neural Network (FFN) in which the information moves in
referred as prospective health and is of great relevance in the only one direction, forward, from the input nodes, through
more general task of personalized medicine. The need for tools the hidden nodes (if there exist) and finally to the output
allowing to perform prediction is a priority. nodes. There are no cycles or loops in the network. Deep
In the recent past EHRs were analyzed using traditional learning algorithms are built upon Artificial Neural Network
machine learning techniques, whereas recently the progress in (ANN) in which a number of interconnected nodes (neurons)
the field of deep learning let to the application of deep learning are organized in a number of different layers. Hidden nodes
techniques to EHRs. are those not belonging to the input or output layers.
The interest from both industries and academia on the use
of deep learning on EHR at different level and with different
specific tasks is testified by the effort devoted by the general
research community on this relevant reseach topic, by the
number of proposed applications and by the general interest on
the topic by public media, premier conferences and prestigious
journals. The amount of literature in this domain is huge and
revise it is out of the scope of the paper.
This paper reports a brief overview of some recently devel-
oped deep learning tools for EHRs.
II. BACKGROUND
a) Electronic Health Records (EHR): The term Fig. 1. NN with 1 input layer, 2 hidden layer and 1 output layer
electronic health record (EHR), or electronic medical record
(EMR), refers to the collection of patient health information
III. T OOLS FOR EHR
in a digital format. They can be categorized in terms of
functionalities: (i) basic EHRs without clinical notes, (ii) basic A. Deepr
EHRs with clinical notes and (iii) comprehensive systems. The paper in [6] presents Deepr, a deep learning system that
EHRs, even in its simple form, provide a rich collection learns how to extract features medical records with the final
of data for the researchers. Data can be shared across the aim of predicting risk for the patient. Deepr is a multi layered
network and may include, as previously specified, a range of architecture based on convolutional neural nets (CNNs). A
information. EHRs were primary designed for internal hospital medical record is a sequence of visits and for each visit there
administrative tasks and many different schemata exist in is a subset of coded diagnosis, lab tests and textual text. A
different structure. Therefore, one of the major challenge young person is expected to have few medical records, whereas
consists in integrating and harmonizing data belonging to an old person is expected to have more medical records.
different institutions. In addition the heterogeneous nature of Medical records are episodic, i.e. they are recorded when
the data types including numerical data, date time objects, patient visit clinics and the time interval between two visit
free text etc poses significant challenges in working with is irregular. Therefore, medical records are irregular at patient
EHRs. level. Anyhow, data are regular at local episode level and these
regularities can be thought of as clinical motifs. Deepr stems
b) Deep Learning: A Neural Network (NN) can be from this observation: the key point is being able to identify
represented by a directed acyclic graph in which the input regular clinical motifs of irregular data.
layer takes in signal vectors and one or multiple hidden layers In order to describe the system we refer to Fig. 2, provided
process the outputs of the previous layer. Deep learning model in [6]. In the following steps are reported from left-to-right.
is the approach aiming at producing an end-to-end system • The top-left box in the figure represents a generic exam-
that learns from raw data and performs specific tasks without ple of a medical record consisting of multiple visits each
manual supervision. A Deep Neural Network presents many having a number of associated coded objects (procedures
more layers and many more nodes in each layer with respect and diagnosis). The risk of the patient is unknown and is
to a neural network. The number of parameters to be tuned represented in the box by a question mark.
increases with respect to those of a neural network and, in • The medical record is transformed into a “sentence” of
addition, they cannot be learned without enough data and multiple phrases (each phrase represents a visit), sepa-
without powerful computers. rated by “special words” that denote the time gap between
There exists different types of deep neural architectures such two visits. This allows to represent a medical record as
as: (i) the Convolutional Neural Network (CNN) in which a sentence containing information of all important events
the connectivity pattern between layers is inspired by the whose length is variable.
organization of the visual cortex; (ii) the Recurrent Neural • The other layers of Deepr constitutes a CNN. At the first,
Network (RNN) where connections between nodes form a words are embedded into a continuous vector space, next,
directed graph along a sequence and (iii) the Feedforward using convolution operators, local motifs are detected.
Fig. 2. Deepr Architercture Overview
B. Deep Patient Fig. 3. Deep Patient: Conceptual framework used to derive the deep
patient representation from EHR data warehouse
The work in [10] proposes the Deep Patient model: an
unsupervised deep learning approach to extract the hierarchical
features and patterns from EHRs data. The paper shows that
unsupervised deep learning can be profitable used for obtain-
ing the underlaying hierarchical features of EHRs datasets.
In order to describe the system we refer to Fig. 3 and Fig.
4 provided in [10]. Figure 3 reports the high level conceptual
framework used to obtain the Deep Patient representation.
EHRs are extracted from the hospital data warehouse, after
they are pre-processed to derive clinically relevant phenotypes
and are grouped in patient vectors (i.e., raw representation, Fig. 4. Deep Patient: Diagram of the unsupervised deep feature
Fig.3A). Each patient is represented by just a single vector or learning pipeline
by a sequence of vectors. The set of vectors, obtained from
all the patients, is used as input of the unsupervised deep
feature learning algorithm that derives high level descriptors model, the disease prediction is evaluated in two different
(Fig.3B). Each patient of the data warehouse is represented clinical tasks: disease classification (i.e., evaluation by disease)
using the same features, in addition this kind of deep represen- and patient disease tagging (i.e., evaluation by patient).
tation can be applied to different clinical tasks (Fig.1C). The The Deep Patient model has been experimented over the
unsupervised deep feature learning pipeline that transforms a Mount Sinai data warehouse which collects emergency room,
row dataset into a deep patient representation is reported in inpatient and outpatient visits.
Fig. 4. The diagram reports multiple layers of neural networks
in which each layer is trained in order to derive a higher- C. Med2Vec
level representation with respect to the output of the previous Med2Vec [7] is a simple and robust algorithm that is able
layer. The last network of the chain outputs the final patient to efficiently learn code and visit level representation by using
representation. This process, that is similar to neuroscience real world EHRs datasets, avoiding the need of expert medical
models of cognition, combines lower-level features into a knowledge. Med2Vec uses a two level neural network for
unified and more compact representation. learning lower dimension representations of clinical concepts
The Deep Patient framework essentially uses stacked de- from large EHRs datasets with over million visits and allows
noising autoencoders (SDA) to learn the data representations to interpret the learned representation.
from multi-domain clinical data. After the construction of the In order to describe the system we refer to the architecture
in Figure 5 provided in [7]. The following description reports
bottom-up steps.