Reading 4 Big Data Projects - Answers
Reading 4 Big Data Projects - Answers
A) feature selection.
B) feature engineering.
C) feature design.
Explanation
Data exploration encompasses exploratory data analysis, feature selection, and feature
engineering.
A) veracity.
B) velocity.
C) variety.
Explanation
Big data is defined as data with high volume, velocity, and variety. Big data often suffers from
low veracity, because it can contain a high percentage of meaningless data.
Explanation
Underfitting describes a machine learning model that is not complex enough to describe the
data it is meant to analyze. An underfit model treats true parameters as noise and fails to
identify the actual patterns and relationships. A model that is overfit (too complex) will tend
to identify spurious relationships in the data. Labelling of input data is related to the use of
supervised or unsupervised machine learning techniques.
A data technician accesses an offsite archive to retrieve data that has been stored
A)
there.
An investor creates a word cloud from financial analysts’ recent research reports
B)
about a company.
C) An analyst gathering data for sentiment analysis determines what sources to use.
Explanation
Data collection (curation) is determining the sources of data to be used (e.g., web scouring,
specific social media sites). Word clouds are a visualization technique. Moving data from a
storage medium to where they are needed is referred to as transfer.
When evaluating the fit of a machine learning algorithm, it is most accurate to state that:
Explanation
Recall (also called sensitivity) is the ratio of correctly predicted positive classes to all actual
positive classes. Precision is the ratio of correctly predicted positive classes to all predicted
positive classes. Accuracy is the percentage of correctly predicted classes out of total
predictions.
In big data analysis, the three primary tasks involved in data exploration are most accurately
described as:
Explanation
Data exploration involves three central tasks: exploratory data analysis, feature selection,
and feature engineering. Exploratory data analysis uses visualizations to observe and
summarize data. Feature selection is where only pertinent features from the dataset are
selected for machine learning model training. Feature engineering is the process of creating
new features by changing or transforming existing features.
An executive describes her company's "low latency, multiple terabyte" requirements for
managing Big Data. To which characteristics of Big Data is the executive referring?
Explanation
Big Data may be characterized by its volume (the amount of data available), velocity (the
speed at which data are communicated), and variety (degrees of structure in which data
exist). "Terabyte" is a measure of volume. "Latency" refers to velocity.
In big data analysis, the most appropriate method of gaining a high-level picture of the
composition of textual content is through the use of a:
A) scatterplot.
B) histogram.
C) word cloud.
Explanation
Word clouds are an effective way to gain a high-level picture of the composition of textual
content. Histograms, box plots, and scatterplots are common techniques for exploring
structured data.
The process of splitting a given text into separate words is best characterized as:
A) tokenization.
B) stemming.
C) bag-of-words.
Explanation
A) 91%.
B) 71%.
C) 81%.
Explanation
Precision, the ratio of correctly predicted positive classes (true positives) to all predicted
positive classes, is calculated as:
In the context of this default classification, high precision would help us avoid the situation
where a bond is incorrectly predicted to default when it actually is not going to default.
Karlsson is especially concerned about the possibility that her model may indicate that a bond
will not default, but then the bond actually defaults. Karlsson decides to use the model's recall
to evaluate this possibility. Based on the data in Exhibit 1, the model's recall is closest to:
A) 83%.
B) 73%.
C) 93%.
Explanation
Recall is useful when the cost of a false negative is high, such as when we predict that a bond
will not default but it actually will. In cases like this, high recall indicates that false negatives
will be minimized.
Karlsson would like to gain a sense of her model's overall performance. In her research,
Karlsson learns about the F1 score, which she hopes will provide a useful measure. Based on
Exhibit 1, Karlsson's model's F1 score is closest to:
A) 72%.
B) 82%.
C) 92%.
Explanation
The model's F1 score, which is the harmonic mean of precision and recall, is calculated as:
Like accuracy, F1 is a measure of overall performance measures that gives equal weight to FP
and FN.
Karlsson also learns of the model measure of accuracy. Based on Exhibit 1, Karlsson's model's
accuracy metric is closest to:
A) 79%.
B) 89%.
C) 69%.
Explanation
The model's accuracy is the percentage of correctly predicted classes out of total predictions.
Model accuracy is calculated as: