Lec3. Outlier Analysis
Lec3. Outlier Analysis
Outlier Analysis
◼ Outlier and Outlier Analysis
◼ Outlier Detection Methods
◼ Statistical Approaches
◼ Proximity-Base Approaches
◼ Clustering-Base Approaches
◼ Classification Approaches
◼ Mining Contextual and Collective Outliers
◼ Outlier Detection in High Dimensional Data
◼ Summary
1
What Are Outliers?
◼ Outlier: A data object that deviates significantly from the normal
objects as if it were generated by a different mechanism
◼ Ex.: Unusual credit card purchase, sports: Michael Jordon, Wayne
Gretzky, ...
◼ Outliers are different from the noise data
◼ Noise is random error or variance in a measured variable
◼ Customer segmentation
◼ Medical analysis
2
Types of Outliers (I)
◼ Three kinds: global, contextual and collective outliers
◼ Global outlier (or point anomaly) Global Outlier
o
◼ Ex. 80 F in Urbana: outlier? (depending on summer or winter?)
3
Types of Outliers (II)
◼ Collective Outliers
◼ A subset of data objects collectively deviate
significantly from the whole data set, even if the
individual data objects may not be outliers
◼ Applications: E.g., intrusion detection: Collective Outlier
◼ When a number of computers keep sending
denial-of-service packages to each other
◼ Detection of collective outliers
◼ Consider not only behavior of individual objects, but also that of
groups of objects
◼ Need to have the background knowledge on the relationship
◼ The border between normal and outlier objects is often a gray area
between normal objects and outliers. It may help hide outliers and
reduce the effectiveness of outlier detection
◼ Understandability
◼ Understand why these are outliers: Justification of the detection
◼ Model normal objects & report those not matching the model as
outliers, or
◼ Model outliers and treat those not matching the model as normal
◼ Challenges
◼ Imbalanced classes, i.e., outliers are rare: Boost the outlier class
◼ Problem 2: Costly since first clustering: but far less outliers than
normal objects
◼ Newer methods: tackle outliers directly
8
Outlier Detection III: Semi-Supervised Methods
◼ Situation: In many applications, the number of labeled data is often
small: Labels could be on outliers only, normal objects only, or both
◼ Semi-supervised outlier detection: Regarded as applications of semi-
supervised learning
◼ If some labeled normal objects are available
◼ Use the labeled examples and the proximate unlabeled objects to
train a model for normal objects
◼ Those not fitting the model of normal objects are detected as outliers
◼ If only some labeled outliers are available, a small number of labeled
outliers many not cover the possible outliers well
◼ To improve the quality of outlier detection, one can get help from
models for normal objects learned from unsupervised methods
9
Outlier Detection (1): Statistical Methods
◼ Statistical methods (also known as model-based methods) assume
that the normal data follow some statistical model (a stochastic model)
◼ The data not following the model are outliers.
◼ Example (right figure): First use Gaussian distribution
to model the normal data
◼ For each object y in region R, estimate gD(y), the
10
Outlier Detection (2): Proximity-Based Methods
◼ An object is an outlier if the nearest neighbors of the object are far
away, i.e., the proximity of the object is significantly deviates from
the proximity of most of the other objects in the same data set
◼ Example (right figure): Model the proximity of an
object using its 3 nearest neighbors
◼ Objects in region R are substantially different
from other objects in the data set.
◼ Thus the objects in R are outliers
◼ The effectiveness of proximity-based methods highly relies on the
proximity measure.
◼ In some applications, proximity or distance measures cannot be
obtained easily.
◼ Often have a difficulty in finding a group of outliers which stay close to
each other
◼ Two major types of proximity-based outlier detection
◼ Distance-based vs. density-based
11
Outlier Detection (3): Clustering-Based Methods
◼ Normal data belong to large and dense clusters, whereas
outliers belong to small or sparse clusters, or do not belong
to any clusters
◼ Example (right figure): two clusters
◼ All points not in R form a large cluster
◼ The two points in R form a tiny cluster,
thus are outliers
◼ Since there are many clustering methods, there are many
clustering-based outlier detection methods as well
◼ Clustering is expensive: straightforward adaption of a
clustering method for outlier detection can be costly and
does not scale up well for large data sets
12
Chapter 12. Outlier Analysis
◼ Outlier and Outlier Analysis
◼ Outlier Detection Methods
◼ Statistical Approaches
◼ Proximity-Base Approaches
◼ Clustering-Base Approaches
◼ Classification Approaches
◼ Mining Contextual and Collective Outliers
◼ Outlier Detection in High Dimensional Data
◼ Summary
13
Statistical Approaches
◼ Statistical approaches assume that the objects in a data set are
generated by a stochastic process (a generative model)
◼ Idea: learn a generative model fitting the given data set, and then
identify the objects in low probability regions of the model as outliers
◼ Methods are divided into two categories: parametric vs. non-parametric
◼ Parametric method
◼ Assumes that the normal data is generated by a parametric
distribution with parameter θ
◼ The probability density function of the parametric distribution f(x, θ)
gives the probability that object x is generated by the distribution
◼ The smaller this value, the more likely x is an outlier
◼ Non-parametric method
◼ Not assume an a-priori statistical model and determine the model
from the input data
◼ Not completely parameter free but consider the number and nature
of the parameters are flexible and not fixed in advance
◼ Examples: histogram and kernel density estimation
14
Parametric Methods I: Detection Univariate
Outliers Based on Normal Distribution
◼ Univariate data: A data set involving only one attribute or variable
◼ Often assume that data are generated from a normal distribution, learn
the parameters from the input data, and identify the points with low
probability as outliers
◼ Ex: Avg. temp.: {24.0, 28.9, 28.9, 29.0, 29.1, 29.1, 29.2, 29.2, 29.3, 29.4}
◼ Use the maximum likelihood method to estimate μ and σ
15
Parametric Methods I: The Grubb’s Test
◼ Univariate outlier detection: The Grubb's test (maximum normed residual
test) ─ another statistical method under normal distribution
◼ For each object x in a data set, compute its z-score: x is an outlier if
16
Parametric Methods II: Detection of
Multivariate Outliers
◼ Multivariate data: A data set involving two or more attributes or
variables
◼ Transform the multivariate outlier detection task into a univariate
outlier detection problem
◼ Method 1. Compute Mahalaobis distance
◼ Let ō be the mean vector for a multivariate data set. Mahalaobis
distance for an object o to ō is MDist(o, ō) = (o – ō )T S –1(o – ō)
where S is the covariance matrix
◼ Use the Grubb's test on this measure to detect outliers
◼ Method 2. Use χ2 –statistic:
◼ where Ei is the mean of the i-dimension among all objects, and n is
the dimensionality
◼ If χ2 –statistic is large, then object oi is an outlier
17
Parametric Methods III: Using Mixture of
Parametric Distributions
◼ Assuming data generated by a normal distribution
could be sometimes overly simplified
◼ Example (right figure): The objects between the two
clusters cannot be captured as outliers since they
are close to the estimated mean
◼ To overcome this problem, assume the normal data is generated by two
normal distributions. For any object o in the data set, the probability that
o is generated by the mixture of the two distributions is given by
where fθ1 and fθ2 are the probability density functions of θ1 and θ2
◼ Then use EM algorithm to learn the parameters μ1, σ1, μ2, σ2 from data
◼ An object o is an outlier if it does not belong to any cluster
18
Non-Parametric Methods: Detection Using Histogram
◼ The model of normal data is learned from the
input data without any a priori structure.
◼ Often makes fewer assumptions about the data,
and thus can be applicable in more scenarios
◼ Outlier detection using histogram:
◼ Figure shows the histogram of purchase amounts in transactions
◼ A transaction in the amount of $7,500 is an outlier, since only 0.2%
transactions have an amount higher than $5,000
◼ Problem: Hard to choose an appropriate bin size for histogram
◼ Too small bin size → normal objects in empty/rare bins, false positive
◼ Too big bin size → outliers in some frequent bins, false negative
◼ Solution: Adopt kernel density estimation to estimate the probability
density distribution of the data. If the estimated density function is high,
the object is likely normal. Otherwise, it is likely an outlier.
19
Chapter 12. Outlier Analysis
◼ Outlier and Outlier Analysis
◼ Outlier Detection Methods
◼ Statistical Approaches
◼ Proximity-Base Approaches
◼ Clustering-Base Approaches
◼ Classification Approaches
◼ Mining Contextual and Collective Outliers
◼ Outlier Detection in High Dimensional Data
◼ Summary
20
Proximity-Based Approaches: Distance-Based vs.
Density-Based Outlier Detection
◼ Intuition: Objects that are far away from the others are
outliers
◼ Assumption of proximity-based approach: The proximity of
an outlier deviates significantly from that of most of the
others in the data set
◼ Two types of proximity-based outlier detection methods
◼ Distance-based outlier detection: An object o is an
outlier if its neighborhood does not have enough other
points
◼ Density-based outlier detection: An object o is an outlier
if its density is relatively much lower than that of its
neighbors
21
Distance-Based Outlier Detection
◼ For each object o, examine the # of other objects in the r-neighborhood
of o, where r is a user-specified distance threshold
◼ An object o is an outlier if most (taking π as a fraction threshold) of
the objects in D are far away from o, i.e., not in the r-neighborhood of o
◼ The lower the local reachability density of o, and the higher the local
reachability density of the kNN of o, the higher LOF
◼ This captures a local outlier whose local density is relatively low
comparing to the local densities of its kNN
25
Chapter 12. Outlier Analysis
◼ Outlier and Outlier Analysis
◼ Outlier Detection Methods
◼ Statistical Approaches
◼ Proximity-Base Approaches
◼ Clustering-Base Approaches
◼ Classification Approaches
◼ Mining Contextual and Collective Outliers
◼ Outlier Detection in High Dimensional Data
◼ Summary
26
Clustering-Based Outlier Detection (1 & 2):
Not belong to any cluster, or far from the closest one
◼ An object is an outlier if (1) it does not belong to any cluster, (2) there is
a large distance between the object and its closest cluster , or (3) it
belongs to a small or sparse cluster
◼ Case I: Not belong to any cluster
◼ Identify animals not part of a flock: Using a density-
◼ Ex. In the figure, o is outlier since its closest large cluster is C1, but the
similarity between o and C1 is small. For any point in C3, its closest
large cluster is C2 but its similarity from C2 is low, plus |C3| = 3 is small
28
Clustering-Based Method: Strength and Weakness
◼ Strength
◼ Detect outliers without requiring any labeled data
◼ Once the cluster are obtained, need only compare any object
31
Classification-Based Method II: Semi-Supervised Learning
◼ Semi-supervised learning: Combining classification-
based and clustering-based methods
◼ Method
◼ Using a clustering-based approach, find a large
cluster, C, and a small cluster, C1
◼ Since some objects in C carry the label “normal”,
treat all objects in C as normal
◼ Use the one-class model of this cluster to identify
normal objects in outlier detection
◼ Since some objects in cluster C1 carry the label
“outlier”, declare all objects in C1 as outliers
◼ Any object that does not fall into the model for C
(such as a) is considered an outlier as well
◼ Comments on classification-based outlier detection methods
◼ Strength: Outlier detection is fast
the training set, but often difficult to obtain representative and high-
quality training data
32
Chapter 12. Outlier Analysis
◼ Outlier and Outlier Analysis
◼ Outlier Detection Methods
◼ Statistical Approaches
◼ Proximity-Base Approaches
◼ Clustering-Base Approaches
◼ Classification Approaches
◼ Mining Contextual and Collective Outliers
◼ Outlier Detection in High Dimensional Data
◼ Summary
33
Mining Contextual Outliers I: Transform into
Conventional Outlier Detection
◼ If the contexts can be clearly identified, transform it to conventional
outlier detection
1. Identify the context of the object using the contextual attributes
2. Calculate the outlier score for the object in the context using a
conventional outlier detection method
◼ Ex. Detect outlier customers in the context of customer groups
◼ Contextual attributes: age group, postal code
◼ Steps: (1) locate c’s context, (2) compare c with the other customers in
the same group, and (3) use a conventional outlier detection method
◼ If the context contains very few customers, generalize contexts
◼ Ex. Learn a mixture model U on the contextual attributes, and
another mixture model V of the data on the behavior attributes
◼ Learn a mapping p(Vi|Uj): the probability that a data object o
belonging to cluster Uj on the contextual attributes is generated by
cluster Vi on the behavior attributes
◼ Outlier score:
34
Mining Contextual Outliers II: Modeling Normal
Behavior with Respect to Contexts
◼ In some applications, one cannot clearly partition the data into contexts
◼ Ex. if a customer suddenly purchased a product that is unrelated to
those she recently browsed, it is unclear how many products
browsed earlier should be considered as the context
◼ Model the “normal” behavior with respect to contexts
◼ Using a training data set, train a model that predicts the expected
behavior attribute values with respect to the contextual attribute
values
◼ An object is a contextual outlier if its behavior attribute values
significantly deviate from the values predicted by the model
◼ Using a prediction model that links the contexts and behavior, these
methods avoid the explicit identification of specific contexts
◼ Methods: A number of classification and prediction techniques can be
used to build such models, such as regression, Markov Models, and
Finite State Automaton
35
Mining Collective Outliers I: On the Set
of “Structured Objects”
◼ Collective outlier if objects as a group deviate
significantly from the entire data
◼ Need to examine the structure of the data set, i.e, the
relationships between multiple data objects
◼ Each of these structures is inherent to its respective type of data
◼ For temporal data (such as time series and sequences), we explore
the structures formed by time, which occur in segments of the time
series or subsequences
◼ For spatial data, explore local areas
◼ Difference from the contextual outlier detection: the structures are often
not explicitly defined, and have to be discovered as part of the outlier
detection process.
◼ Collective outlier detection methods: two categories
◼ Reduce the problem to conventional outlier detection
application dependent
◼ The computational cost is often high due to the sophisticated
mining process
37
Chapter 12. Outlier Analysis
◼ Outlier and Outlier Analysis
◼ Outlier Detection Methods
◼ Statistical Approaches
◼ Proximity-Base Approaches
◼ Clustering-Base Approaches
◼ Classification Approaches
◼ Mining Contextual and Collective Outliers
◼ Outlier Detection in High Dimensional Data
◼ Summary
38
Challenges for Outlier Detection in High-
Dimensional Data
◼ Interpretation of outliers
◼ Detecting outliers without saying why they are outliers is not very
39
Approach I: Extending Conventional Outlier
Detection
◼ Method 1: Detect outliers in the full space, e.g., HilOut Algorithm
◼ Find distance-based outliers, but use the ranks of distance instead of
◼ The more negative, the sparser C is and the more likely the
44
References (I)
◼ B. Abraham and G.E.P. Box. Bayesian analysis of some outlier problems in time series. Biometrika, 66:229–248,
1979.
◼ M. Agyemang, K. Barker, and R. Alhajj. A comprehensive survey of numeric and symbolic outlier mining
techniques. Intell. Data Anal., 10:521–538, 2006.
◼ F. J. Anscombe and I. Guttman. Rejection of outliers. Technometrics, 2:123–147, 1960.
◼ D. Agarwal. Detecting anomalies in cross-classified streams: a bayesian approach. Knowl. Inf. Syst., 11:29–44,
2006.
◼ F. Angiulli and C. Pizzuti. Outlier mining in large high-dimensional data sets. TKDE, 2005.
◼ C. C. Aggarwal and P. S. Yu. Outlier detection for high dimensional data. SIGMOD’01
◼ R.J. Beckman and R.D. Cook. Outlier...s. Technometrics, 25:119–149, 1983.
◼ I. Ben-Gal. Outlier detection. In Maimon O. and Rockach L. (eds.) Data Mining and Knowledge Discovery
Handbook: A Complete Guide for Practitioners and Researchers, Kluwer Academic, 2005.
◼ M. M. Breunig, H.-P. Kriegel, R. Ng, and J. Sander. LOF: Identifying density-based local outliers. SIGMOD’00
◼ D. Barbar´a, Y. Li, J. Couto, J.-L. Lin, and S. Jajodia. Bootstrapping a data mining intrusion detection system.
SAC’03
◼ Z. A. Bakar, R. Mohemad, A. Ahmad, and M. M. Deris. A comparative study for outlier detection techniques in
data mining. IEEE Conf. on Cybernetics and Intelligent Systems, 2006.
◼ S. D. Bay and M. Schwabacher. Mining distance-based outliers in near linear time with randomization and a
simple pruning rule. KDD’03
◼ D. Barbara, N. Wu, and S. Jajodia. Detecting novel network intrusion using bayesian estimators. SDM’01
◼ V. Chandola, A. Banerjee, and V. Kumar. Anomaly detection: A survey. ACM Computing Surveys, 41:1–58,
2009.
◼ D. Dasgupta and N.S. Majumdar. Anomaly detection in multidimensional data using negative selection
algorithm. In CEC’02
References (2)
◼ E. Eskin, A. Arnold, M. Prerau, L. Portnoy, and S. Stolfo. A geometric framework for unsupervised anomaly
detection: Detecting intrusions in unlabeled data. In Proc. 2002 Int. Conf. of Data Mining for Security
Applications, 2002.
◼ E. Eskin. Anomaly detection over noisy data using learned probability distributions. ICML’00
◼ T. Fawcett and F. Provost. Adaptive fraud detection. Data Mining and Knowledge Discovery, 1:291–316, 1997.
◼ V. J. Hodge and J. Austin. A survey of outlier detection methdologies. Artif. Intell. Rev., 22:85–126, 2004.
◼ D. M. Hawkins. Identification of Outliers. Chapman and Hall, London, 1980.
◼ Z. He, X. Xu, and S. Deng. Discovering cluster-based local outliers. Pattern Recogn. Lett., 24, June, 2003.
◼ W. Jin, K. H. Tung, and J. Han. Mining top-n local outliers in large databases. KDD’01
◼ W. Jin, A. K. H. Tung, J. Han, and W. Wang. Ranking outliers using symmetric neighborhood relationship.
PAKDD’06
◼ E. Knorr and R. Ng. A unified notion of outliers: Properties and computation. KDD’97
◼ E. Knorr and R. Ng. Algorithms for mining distance-based outliers in large datasets. VLDB’98
◼ E. M. Knorr, R. T. Ng, and V. Tucakov. Distance-based outliers: Algorithms and applications. VLDB J., 8:237–
253, 2000.
◼ H.-P. Kriegel, M. Schubert, and A. Zimek. Angle-based outlier detection in high-dimensional data. KDD’08
◼ M. Markou and S. Singh. Novelty detection: A review—part 1: Statistical approaches. Signal Process., 83:2481–
2497, 2003.
◼ M. Markou and S. Singh. Novelty detection: A review—part 2: Neural network based approaches. Signal
Process., 83:2499–2521, 2003.
◼ C. C. Noble and D. J. Cook. Graph-based anomaly detection. KDD’03
References (3)
◼ S. Papadimitriou, H. Kitagawa, P. B. Gibbons, and C. Faloutsos. Loci: Fast outlier detection using the local
correlation integral. ICDE’03
◼ A. Patcha and J.-M. Park. An overview of anomaly detection techniques: Existing solutions and latest
technological trends. Comput. Netw., 51, 2007.
◼ X. Song, M. Wu, C. Jermaine, and S. Ranka. Conditional anomaly detection. IEEE Trans. on Knowl. and Data
Eng., 19, 2007.
◼ Y. Tao, X. Xiao, and S. Zhou. Mining distance-based outliers from large databases in any metric space. KDD’06
◼ N. Ye and Q. Chen. An anomaly detection technique based on a chi-square statistic for detecting intrusions
into information systems. Quality and Reliability Engineering International, 17:105–112, 2001.
◼ B.-K. Yi, N. Sidiropoulos, T. Johnson, H. V. Jagadish, C. Faloutsos, and A. Biliris. Online data mining for co-
evolving time sequences. ICDE’00
Un-Used Slides
48
Outlier Discovery:
Statistical Approaches
◼ data distribution
◼ Drawbacks
49
Outlier Discovery: Distance-Based Approach
◼ Introduced to counter the main limitations imposed by
statistical methods
◼ We need multi-dimensional analysis without knowing
data distribution
◼ Distance-based outlier: A DB(p, D)-outlier is an object O in
a dataset T such that at least a fraction p of the objects in T
lies at a distance greater than D from O
◼ Algorithms for mining distance-based outliers [Knorr & Ng,
VLDB’98]
◼ Index-based algorithm
◼ Nested-loop algorithm
◼ Cell-based algorithm
50
Density-Based Local
Outlier Detection
52
References (1)
◼ B. Abraham and G.E.P. Box. Bayesian analysis of some outlier problems in time series. Biometrika,
1979.
◼ Malik Agyemang, Ken Barker, and Rada Alhajj. A comprehensive survey of numeric and symbolic
outlier mining techniques. Intell. Data Anal., 2006.
◼ Deepak Agarwal. Detecting anomalies in cross-classied streams: a bayesian approach. Knowl. Inf.
Syst., 2006.
◼ C. C. Aggarwal and P. S. Yu. Outlier detection for high dimensional data. SIGMOD'01.
◼ M. M. Breunig, H.-P. Kriegel, R. T. Ng, and J. Sander. Optics-of: Identifying local outliers. PKDD '99
◼ M. M. Breunig, H.-P. Kriegel, R. Ng, and J. Sander. LOF: Identifying density-based local outliers.
SIGMOD'00.
◼ V. Chandola, A. Banerjee, and V. Kumar. Anomaly detection: A survey. ACM Comput. Surv., 2009.
◼ D. Dasgupta and N.S. Majumdar. Anomaly detection in multidimensional data using negative
selection algorithm. Computational Intelligence, 2002.
◼ E. Eskin, A. Arnold, M. Prerau, L. Portnoy, and S. Stolfo. A geometric framework for unsupervised
anomaly detection: Detecting intrusions in unlabeled data. In Proc. 2002 Int. Conf. of Data Mining
for Security Applications, 2002.
◼ E. Eskin. Anomaly detection over noisy data using learned probability distributions. ICML’00.
◼ T. Fawcett and F. Provost. Adaptive fraud detection. Data Mining and Knowledge Discovery, 1997.
◼ R. Fujimaki, T. Yairi, and K. Machida. An approach to spacecraft anomaly detection problem using
kernel feature space. KDD '05
◼ F. E. Grubbs. Procedures for detecting outlying observations in samples. Technometrics, 1969.
53
References (2)
◼ V. Hodge and J. Austin. A survey of outlier detection methodologies. Artif. Intell. Rev., 2004.
◼ Douglas M Hawkins. Identification of Outliers. Chapman and Hall, 1980.
◼ P. S. Horn, L. Feng, Y. Li, and A. J. Pesce. Effect of Outliers and Nonhealthy Individuals on
Reference Interval Estimation. Clin Chem, 2001.
◼ W. Jin, A. K. H. Tung, J. Han, and W. Wang. Ranking outliers using symmetric neighborhood
relationship. PAKDD'06
◼ E. Knorr and R. Ng. Algorithms for mining distance-based outliers in large datasets. VLDB’98
◼ M. Markou and S. Singh.. Novelty detection: a review| part 1: statistical approaches. Signal
Process., 83(12), 2003.
◼ M. Markou and S. Singh. Novelty detection: a review| part 2: neural network based approaches.
Signal Process., 83(12), 2003.
◼ S. Papadimitriou, H. Kitagawa, P. B. Gibbons, and C. Faloutsos. Loci: Fast outlier detection using
the local correlation integral. ICDE'03.
◼ A. Patcha and J.-M. Park. An overview of anomaly detection techniques: Existing solutions and
latest technological trends. Comput. Netw., 51(12):3448{3470, 2007.
◼ W. Stefansky. Rejecting outliers in factorial designs. Technometrics, 14(2):469{479, 1972.
◼ X. Song, M. Wu, C. Jermaine, and S. Ranka. Conditional anomaly detection. IEEE Trans. on Knowl.
and Data Eng., 19(5):631{645, 2007.
◼ Y. Tao, X. Xiao, and S. Zhou. Mining distance-based outliers from large databases in any metric
space. KDD '06:
◼ N. Ye and Q. Chen. An anomaly detection technique based on a chi-square statistic for detecting
intrusions into information systems. Quality and Reliability Engineering International, 2001.
54