0% found this document useful (0 votes)
27 views45 pages

Tsa - Time Series Analysis

Uploaded by

krishnatanna41
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
27 views45 pages

Tsa - Time Series Analysis

Uploaded by

krishnatanna41
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 45

Introduction to Time Series and Forecasting

A time series is a sequence of data points collected or recorded at regular time intervals. Time series
analysis involves studying these data points to understand the underlying patterns, trends, and relationships.
Forecasting uses these patterns to predict future values.
Time series analysis is widely used in various fields, including finance (stock prices), economics (GDP
growth), meteorology (weather prediction), healthcare (disease trends), and business (sales forecasting).
Key concepts:
• Temporal dependence: Data points are correlated with their past values.
• Stationarity: A time series is stationary if its statistical properties (mean, variance) remain constant
over time.
• Seasonality & Trend: Some time series exhibit repeating patterns (seasonality) or long-term
directional movement (trend).

Different Types of Data in Time Series


Time series data can be categorized based on its characteristics:
1. Stationary vs. Non-Stationary Data:
o Stationary: Statistical properties remain constant over time (e.g., white noise).
o Non-stationary: Trends, seasonality, or changing variance over time.
2. Univariate vs. Multivariate Data:
o Univariate: A single variable measured over time (e.g., monthly sales).
o Multivariate: Multiple interdependent variables measured over time (e.g., sales and
advertising expenditure).
3. Discrete vs. Continuous Data:
o Discrete: Data is recorded at specific intervals (e.g., daily stock prices).
o Continuous: Data is collected continuously (e.g., real-time temperature readings).
4. Deterministic vs. Stochastic Data:
o Deterministic: Can be predicted with certainty (e.g., calendar dates).
o Stochastic: Includes random fluctuations (e.g., stock market returns).

Internal Structures of Time Series


Time series data exhibits various internal structures:
Trend (T): The long-term increase or decrease in the data (e.g., GDP growth). If a trend is increasing, the
graph moves upwards. If it is decreasing, it moves downwards.
Example:
• The world’s average temperature has been rising over the last century.
• The stock market index has mostly gone up over many years.
• A company’s sales may increase steadily due to business growth.
Seasonality (S): Periodic patterns that repeat at fixed intervals (e.g., monthly sales increase during festive
seasons). Seasonality refers to patterns that repeat at fixed intervals (daily, monthly, yearly, etc.). The graph
will show waves or spikes at regular intervals.
Example:
• Retail sales increase every December due to holiday shopping.
• Electricity bills are higher in summer due to air conditioning use.
• Crop production changes with the seasons (e.g., wheat is harvested in winter).

Cyclic Patterns (C): Fluctuations occurring at irregular intervals (e.g., business cycles). Cyclic patterns are
long-term changes in data that do not happen at fixed time intervals. These cycles are often influenced by
economic or social factors. A cycle looks like a wave but does not follow a fixed pattern like seasonality.
Example:
• The economy booms and crashes every few years.
• Real estate prices go up and down over time.
• The stock market experiences bull (rising) and bear (falling) phases.

Irregular Component (I): Random variations or noise in data (e.g., sudden market crashes). The irregular
component includes random, unpredictable events that affect the data. The graph will show sudden spikes or
drops that do not follow any pattern.
Example:
• A sudden stock market crash due to an economic crisis.
• A natural disaster like an earthquake affecting agricultural production.
• The COVID-19 pandemic causing a drop in airline travel.

A time series can be expressed as:

Models for Time Series Analysis


Several statistical and machine learning models are used for time series forecasting:
1. Classical Statistical Models:
o Moving Average (MA): Averages past observations to smooth the series.
o Autoregressive (AR): Uses past values to predict future ones.
o ARMA (AutoRegressive Moving Average): Combines AR and MA.
o ARIMA (AutoRegressive Integrated Moving Average): Extends ARMA by differencing
to handle non-stationary data.
2. Machine Learning Models:
o Regression Models (Linear, Lasso, Ridge)
o Decision Trees & Random Forests
o Neural Networks (RNN, LSTM, GRU)
3. Deep Learning Models:
o LSTM (Long Short-Term Memory) Networks: Handles sequential dependencies well.
o Transformers (e.g., Time Series Transformer)

Autocorrelation and Partial Autocorrelation


• Autocorrelation: Measures the correlation of a time series with its past values. It helps determine
how past values influence future ones. Autocorrelation measures how much a time series value at one
point in time is related to previous values (also called lags). It helps us identify repeating patterns in the data.
In other words, it tells us:
o If the past values of a time series have an effect on the present value.
o How strong the relationship is between observations at different time intervals.

1. If ACF(k) is high, it means the data points at time t and t−k are strongly related.
2. ACF helps identify seasonality → If ACF values are high at specific lags (e.g., every 12
months), it suggests a yearly pattern.
3. The ACF plot (Correlogram) shows autocorrelation values for different lags.

• Partial Autocorrelation: Measures the correlation between a time series and a lagged version of
itself after removing the influence of other lags. Partial Autocorrelation measures the relationship between
a time series value and its lagged values, removing the effects of intermediate values. In simple terms,
PACF tells us:
o How much influence a past value has on the present value, ignoring the effects of other
past values.
o Helps determine the correct number of lagged values (p) for an AR model in forecasting.

1. PACF removes the indirect influence of previous lags, so it only shows the direct effect of
past values on the present.
2. The PACF plot helps choose the AR (AutoRegressive) model order for time series
forecasting.
3. If PACF is significant at lag k but drops off after, it suggests that an AR(k) model is suitable.

Plots Used:
• Autocorrelation Function (ACF): Shows correlation for different lags.
• Partial Autocorrelation Function (PACF): Helps determine the order of AR and MA models.

Examples of Time Series


• Stock Market Prices: Daily closing prices of a stock.
• Weather Data: Monthly average temperature readings.
• Economic Indicators: GDP, inflation rates.
• Sales Data: Monthly sales revenue.
• Electricity Demand: Hourly power consumption.

Nature and Uses of Forecasting


Nature of Forecasting:
• Predictive Analytics: Anticipates future trends based on past patterns.
• Quantitative & Qualitative Methods: Uses mathematical models or expert judgment.
• Short-term vs. Long-term Forecasting: Short-term for operations, long-term for strategic planning.
Uses of Forecasting:
• Business: Demand forecasting for inventory management.
• Finance: Predicting stock price movements.
• Healthcare: Forecasting disease outbreaks.
• Weather Prediction: Estimating rainfall or temperature.
• Economics: Forecasting GDP growth.
Forecasting Process
A structured forecasting process includes the following steps:
1. Define the Objective: Determine the purpose of forecasting.
2. Collect Data: Gather relevant historical data.
3. Preprocess Data: Handle missing values, remove noise.
4. Identify Patterns: Detect trends, seasonality, and cycles.
5. Choose a Model: Select an appropriate forecasting method.
6. Evaluate the Model: Assess accuracy using metrics like RMSE, MAE.
7. Generate Forecasts: Use the model to make predictions.
8. Monitor and Update: Adjust the model based on real-time performance.

Data for Forecasting


Data quality is crucial for accurate forecasting. Accurate forecasting depends on high-quality data. If the
data is incomplete, inconsistent, or inaccurate, predictions may be unreliable. The required data includes:
• Historical Data: Past records of the variable being predicted. Example: A retail store uses five years
of sales data to forecast next year's revenue.
• Seasonal Data: Information about periodic fluctuations that repeat over time. Example: Ice cream
sales increase in summer and drop in winter due to seasonality.
• Economic Indicators: Macroeconomic factors influencing trends, such as GDP, inflation, and
interest rates. Example: High inflation may reduce consumer spending, affecting sales forecasts.
• Market Trends: Insights into consumer behaviour, demand patterns, and industry trends. Example:
A company analyses e-commerce sales growth before expanding online.
• Real-Time Data: Live updates from sensors, social media, and online activity to capture rapid
changes. Example: A stock market prediction model uses real-time trading data to adjust forecasts.

Resources for Forecasting


Several tools and resources are available for time series forecasting:
1. Software & Libraries:
o Python: Pandas, NumPy, Statsmodels, Scikit-learn, TensorFlow
o R: forecast, tseries, ggplot2
o Excel: Built-in forecasting functions
2. Datasets:
o Kaggle: Time series datasets.
o UCI Machine Learning Repository.
o Government databases (World Bank, IMF).
3. Online Courses:
o Coursera, Udemy, edX.
4. Books:
o "Forecasting: Principles and Practice" by Rob J. Hyndman.
o "Time Series Analysis and Its Applications" by Robert H. Shumway.

A company records its quarterly sales (in million units) over five years:
[50, 55, 70, 90, 52, 57, 75, 95, 54, 60, 80, 100, 56, 62, 85, 110, 58, 65, 90, 120]
Identifying Trend
A trend refers to a consistent increase or decrease in values over time.
• Observing the data, we see that sales are increasing from 50 to 120 over five years.
• The growth is gradual (small increments per quarter).
• Thus, a positive trend exists.
Identifying Seasonality
Seasonality refers to patterns repeating at regular intervals.

Thus, the dataset shows both a trend and seasonality

Q.2 Given monthly sales: [120, 135, 150, 160, 175, 190]
Calculate the 3-month moving average forecast for the next month.
Compute Moving Averages
A 3-month moving average is calculated as:
• Forecast for the 7th month: 188.3
• Actual sales = 200, so the forecast underestimated by 11.7 units.

Given time series:


[200, 250, 300, 350, 400, 450]
Find Lag-1 autocorrelation r1r_1r1 using:
Handling Missing Values in a Time Series Dataset
We are given a dataset with missing values:[1000, 1100, 1200, ?, 1400, ?, 1600, 1700, ?, 1900]
We need to estimate the missing values using linear interpolation, which assumes a smooth and consistent
trend in the data.
completed dataset:[1000, 1100, 1200, 1300, 1400, 1500, 1600, 1700, 1800, 1900]

Types of Graphical Displays in Time Series Analysis


Graphical displays help in visualizing time series data, identifying patterns, and selecting appropriate
forecasting models. Following are the types of graphical displays, their purpose, and when to use them.

1. Line Charts (Time Series Plots)


Why Use?
• Helps to visualize trends, seasonality, and fluctuations over time.
• Useful for spotting long-term growth, decline, or cyclic behaviors.
• Allows comparison of multiple time series.
When to Use?
• While analyzing sales growth, stock prices, temperature trends, or any data recorded over time.
• While identifying patterns like trends (increasing/decreasing) or seasonality (recurring patterns).
• While detecting sudden shifts, anomalies, or unusual spikes.
Example:
• Plotting monthly sales data of a company for the last 5 years.
• Observing climate change trends using global temperature data.

2. Histograms (Frequency Distribution Plots)


Why Use?
• Shows the distribution of data points (e.g., whether the data is normal, skewed, or bimodal).
• Helps to detect outliers and data concentration in specific ranges.
• Useful in probabilistic forecasting by understanding past occurrences.
When to Use?
• When examining the distribution of errors in a forecasting model.
• When analyzing stock market return distributions to assess risk.
• When checking if data follows a normal distribution before using statistical methods.
Example:
• Analyzing weekly demand distribution for a product.
• Checking distribution of monthly rainfall to determine variation.

3. Box Plots (Detecting Outliers and Variability)


Why Use?
• Summarizes key statistics like median, quartiles, and outliers in a dataset.
• Useful for comparing multiple time series side by side.
• Highlights extreme variations (outliers) which may affect forecasts.
When to Use?
• While comparing sales data across multiple years to see variability.
• While identifying outliers in financial data like stock market fluctuations.
• While checking for seasonal variation and anomalies in demand forecasting.
Example:
• A box plot of monthly electricity consumption can show seasonal variations.
• Identifying sudden high or low spikes in air pollution data.

4. Scatter Plots (Exploring Relationships Between Variables)


Why Use?
• Determines the correlation between two variables.
• Helps to see if one variable influences another (e.g., sales vs. advertising spend).
• Used in regression analysis to check relationships.
When to Use?
• While studying how economic indicators affect stock prices.
• While analyzing marketing spend vs. sales revenue to check impact.
• While checking rainfall vs. crop yield for agricultural forecasting.
Example:
• A scatter plot of advertising budget vs. sales can show how marketing impacts revenue.
• A scatter plot of temperature vs. energy consumption can reveal a positive correlation.

5. Lag Plots
Why Use?
• Determines if a time series has a pattern or is random.
• Helps identify autocorrelation (how past values influence future values).
• Assists in choosing ARIMA models for forecasting.
When to Use?
• While checking if today’s stock price is related to yesterday’s price.
• While analysing seasonal effects in sales
• While selecting appropriate forecasting models like ARIMA that rely on autocorrelation.
Example:
• A lag plot for monthly airline passenger data can confirm seasonality.
• A lag plot of daily COVID-19 cases can help in understanding infection trends.

Time Series Plots


A Time Series Plot is a graph where:
• The x-axis represents time.
• The y-axis represents the variable of interest.
Key Features of a Time Series Plot:
1. Trend: Long-term increase or decrease in data.
2. Seasonality: Recurring fluctuations at fixed intervals.
3. Cyclical Component: Patterns occurring over irregular periods.
4. Irregular Component: Unpredictable variations (random noise).
Example:
If we plot monthly sales data for a retail store, we may observe:
• Trend: A general increase in sales over years.
• Seasonality: Sales spike in December (holiday season).
• Irregular component: Sudden drop due to external factors (e.g., recession).
Plotting Smoothed Data
Smoothed data removes short-term fluctuations to highlight underlying trends.
Smoothing Techniques:
1. Moving Average (MA): Averages past values to smooth fluctuations.
o Formula:

2. Exponential Smoothing (ES): Gives more weight to recent observations.


o Formula:

Where α (smoothing parameter) controls the influence of past data.


3. Loess Smoothing: Uses local regression to smooth fluctuations.
4. Kernel Smoothing: Weighted averaging method for trend detection.
Why Use Smoothing?
o Removes short-term noise.
o Helps detect long-term trends.
o Improves forecasting accuracy.
Data Transformations and Adjustments
Transformations help in stabilizing variance, making data normal, and improving model performance.

These transformations are essential when working with time series data, as they help:
1. Stabilize variance (Log Transformation)
2. Remove trends and stationarize data (Differencing)
3. Make data normal (Box-Cox Transformation)
4. Scale data for machine learning models (Normalization)
Why Use Transformations?
• Helps remove heteroscedasticity (changing variance).
• Improves model accuracy.
• Makes forecasting models more effective.

General Approach to Time Series Modeling and Forecasting


The forecasting process involves multiple steps:
Step 1: Understand the Data
• Visualize the data using time series plots.
• Check for trends, seasonality, and outliers.
Step 2: Make Data Stationary
• Apply differencing if trend exists.
• Apply log transformation for varying variance.
Step 3: Choose an Appropriate Model
• Moving Average (MA): Smoothens short-term fluctuations.
• Exponential Smoothing (ETS): Suitable for trend and seasonality.
• ARIMA (AutoRegressive Integrated Moving Average): Best for stationary time series.
• LSTM (Deep Learning Model): For complex patterns in large datasets.
Step 4: Evaluate the Model
• Use error metrics (RMSE, MAE, MAPE) to assess accuracy.
Step 5: Generate Forecasts
• Apply the trained model to predict future values.

Evaluating and Monitoring Forecasting Model Performance


Once a forecasting model is built, its performance must be monitored.
Why Monitor Performance?
• Helps detect errors in forecasts.
• Ensures model does not overfit or underfit.
• Adjusts model parameters based on real-time data.
Problem:
A company tracks the monthly revenue (in $1000s) for five years. The data follows a pattern:
• Trend Component (T): Revenue increases by $2,000 each year.
• Seasonal Component (S): Revenue is 10% higher in December and 10% lower in June.
• Cyclic Component (C): Every 3 years, the revenue drops by $5,000.
• Irregular Component (I): Due to a pandemic, revenue in April 2022 dropped by $7,000.
Calculate the total revenue for:
1. December 2023
2. June 2022
3. April 2022
Exponential Smoothing in Time Series Forecasting
Exponential smoothing is a popular time series forecasting method that assigns exponentially decreasing
weights to past observations, giving higher importance to recent data points. This method is widely used in
forecasting demand, stock prices, weather patterns, and economic indicators.
The key idea behind exponential smoothing is to generate a smoothed version of the time series, which helps
in making short-term predictions by minimizing fluctuations caused by random variations.
• First-order exponential smoothing is best for stationary time series.
• Second-order exponential smoothing is better for trending time series.
• Choosing the right smoothing constants α and β is crucial for accurate forecasting.
• For data with seasonality, Triple Exponential Smoothing (Holt-Winters Method) can be used.
Interpretation:
• The smoothed values adjust gradually towards actual sales.
• The higher the smoothing constant α\alphaα, the more weight is given to recent data.

Second-Order Exponential Smoothing (Double Exponential Smoothing)


Regression Models for General Time Series Data
When working with time-dependent data, we use time series regression models to understand patterns, trends, and
dependencies over time. Unlike simple regression models, time series models take into account the sequential nature
of the data.
The main types of time series regression models include:
1. Autoregressive (AR) Models
2. Moving Average (MA) Models
3. Autoregressive Moving Average (ARMA) Models
4. Autoregressive Integrated Moving Average (ARIMA) Models
Comparison
Introduction to Linear Regression
Linear regression is a fundamental statistical technique used for modeling relationships between variables. It helps in
predicting outcomes based on input variables. The main objective is to find the best-fitting line that represents the
relationship between dependent and independent variables.
• Dependent Variable (Y): The variable we want to predict.
• Independent Variable (X): The predictor variable.
• Linear Relationship: The relationship between X and Y is represented as:

Least Squares Estimation in Linear Regression Models

Statistical Inference in Linear Regression


Statistical inference helps us determine how well our model fits the data.
• Hypothesis Testing: Tests if the relationship between X and Y is statistically significant.
• p-value: If p-value < 0.05, the relationship is significant.
• Confidence Intervals: Used to estimate the true value of regression coefficients.
• R-squared (): Measures the proportion of variance explained by the model. Higher means a better fit.

Hypothesis Testing in Regression


In linear regression, hypothesis testing determines whether an independent variable significantly impacts the
dependent variable.
p-Value in Regression

Confidence Interval
R-squared (R2)

ARMA Model
The Autoregressive Moving Average (ARMA) model is a widely used statistical model for analyzing and
forecasting stationary time series data. It combines two components:
1. Autoregressive (AR) Model – Uses past values of the time series to predict future values.
2. Moving Average (MA) Model – Uses past error terms (shocks) to improve predictions.
Since ARMA assumes stationarity, it is suitable for time series data with no trend or seasonal components. If the
data is non-stationary, it must be transformed using differencing before applying ARMA.
Key Applications of ARMA Model
• Stock Market Prediction – Forecasting short-term price movements.
• Economic Forecasting – Modeling GDP growth rates and inflation.
• Signal Processing – Filtering noise from signals.
• Weather Forecasting – Analyzing temperature variations.

Component of ARMA

Advantages
• Works well for stationary time series.
• Effective for short-term forecasting.
• Less computationally intensive compared to deep learning models.
Limitations
• Cannot handle non-stationary data – Requires transformation.
• Cannot capture seasonality – Use SARIMA instead for seasonal data.
• Model selection is manual – Requires tuning p and q based on ACF/PACF plots.
ARIMA Model
The Autoregressive Integrated Moving Average (ARIMA) model is used for time series forecasting, particularly
for non-stationary data. ARIMA extends the Autoregressive Moving Average (ARMA) model by introducing an
Integrated (I) component, which helps in making the series stationary through differencing.
ARIMA is suitable for datasets with trends but no seasonality. If seasonality exists, we use the Seasonal ARIMA
(SARIMA) model.
Key Applications of ARIMA Model
• Stock Market Forecasting – Predicting price trends.
• Economic Data Modeling – Forecasting GDP, inflation rates.
• Sales Forecasting – Estimating future product demand.
• Weather Forecasting – Predicting temperature variations.

Components of ARIMA Model

Advantages
• Handles Trend – Unlike ARMA, ARIMA can model non-stationary data.
• Good for Short-Term Forecasting – Works well for economic and sales forecasting.
• Widely Used – Many industries rely on ARIMA models.
Limitations
• Does Not Handle Seasonality – Requires SARIMA for seasonal data.
• Sensitive to Parameter Selection – Requires tuning of p,d,q.
• Computationally Intensive – Larger datasets require more processing power.

SARIMA Model
The Seasonal Autoregressive Integrated Moving Average (SARIMA) model is an extension of the ARIMA model
designed to handle seasonal time series data. While ARIMA works well for non-seasonal time series, it does not
explicitly model seasonality. SARIMA addresses this limitation by incorporating seasonal autoregressive, seasonal
differencing, and seasonal moving average components.
SARIMA is particularly useful for datasets with patterns that repeat periodically, such as:
• Monthly sales data (retail sales, airline passengers).
• Daily temperature data (seasonal climate changes).
• Quarterly GDP growth (economic trends).

Mathematical Representation of SARIMA Model

Advantages
• Handles both trend and seasonality effectively.
• Works well with time series data that exhibits seasonal fluctuations.
• Can be fine-tuned using ACF and PACF plots.
Limitations
• Requires manual parameter tuning
• Computationally expensive for long time series datasets.
• May not work well for data with sudden structural changes
Multivariate time series

A multivariate time series (MTS) is a collection of multiple time-dependent variables observed over time. Unlike
univariate time series, which analyze only one variable, MTS models capture the relationships among multiple
variables, helping improve forecasting accuracy and understanding interactions between variables.
Examples of Multivariate Time Series:
• Economics: GDP, inflation rate, and unemployment rate over time.
• Finance: Stock prices, trading volume, and interest rates.
• Weather Forecasting: Temperature, humidity, and wind speed.
• Healthcare: Blood pressure, heart rate, and oxygen levels over time.

Components of Multivariate Time Series


Just like univariate time series, MTS consists of the following components:
1. Trend: Long-term upward or downward movement.
2. Seasonality: Repeating patterns at regular intervals (e.g., quarterly sales).
3. Cyclic Patterns: Irregular fluctuations not related to seasonality.
4. Noise (Residuals): Random variations in the data.

VARIMA Model
The Vector Autoregressive Integrated Moving Average (VARIMA) model is an extension of the VAR (Vector
AutoRegression) and VARMA (Vector AutoRegressive Moving Average) models that accounts for non-
stationarity in multivariate time series data.
• VARIMA models multiple time-dependent variables that influence each other over time.
• It includes differencing (Integrated - I component) to handle non-stationary data.
• It combines autoregressive (AR) and moving average (MA) components for multivariate data analysis.
Example Use Cases of VARIMA:
• Economics: Forecasting GDP, interest rates, and inflation together.
• Finance: Predicting stock market trends using stock prices, exchange rates, and bond yields.
• Climate Science: Studying temperature, humidity, and air pressure trends.
• Energy Forecasting: Predicting electricity consumption based on past consumption, weather data, and
industrial demand.
Components of VARIMA
VARIMA is an extension of the VARMA model with an additional integration (I) component for handling non-
stationary data.
A VARIMA (p, d, q) model consists of:
• p (AutoRegression - AR): Number of past values (lags) used for prediction.
• d (Integration - I): Number of times the series is differenced to make it stationary.
• q (Moving Average - MA): Number of past error terms used to improve predictions.

Steps to Build a VARIMA Model


Step 1: Check for Stationarity
A time series is stationary if its statistical properties (mean, variance, autocorrelation) do not change over time.
• Use time series plot to check for trends.
• Perform test to check stationarity.
• If non-stationary, apply differencing until it becomes stationary.
Step 2: Identify the Order (p, d, q)
• ACF plot helps in selecting q (MA order).
• PACF plot helps in selecting p (AR order).
• Check the number of times differencing is applied (d).
Step 3: Fit the VARIMA Model
• Choose appropriate p, d, q values.
• Train the model on historical data.
• Use statistical software like Python (statsmodels), R, or MATLAB to estimate parameters.
Step 4: Model Evaluation
• Use AIC and BIC to select the best model.
• Check errors to ensure they are normally distributed and uncorrelated.
Step 5: Forecasting
• Once the model is trained, use it to make future predictions.
• Evaluate forecast accuracy using Mean Squared Error (MSE) and Root Mean Squared Error (RMSE).
Advantages of VARIMA
1. Captures relationships between multiple time series (e.g., GDP and inflation).
2. Handles non-stationary data by incorporating differencing.
3. Useful for economic and financial forecasting.
4. More flexible than univariate models (like ARIMA).
Limitations of VARIMA
1. Computationally expensive for high-dimensional data.
2. Requires large datasets for accurate parameter estimation.
3. Sensitive to missing values and outliers.

You might also like