Finance
Finance
A
Lab Report
On
“Advance Research”
Session 2022-24
Submitted by
S.Purna Chandra Rahul Rao
Roll No.: - 500107622059
Enrollment No.: - CC0595
MBA II Semester (Section- ‘C’)
2022-2024
DEPARTMENT OF MANAGEMENT
BHILAI INSTITUTE OF TECHNOLOGY
(Seth Balkrishan Memorial), (Accredited by NBA – AICTE)
(An ISO 9001:2000 Certified Institute), Bhilai House, Durg – 491001 (C.G.)
Appendix – II
A
Lab Report
On
“Advance Research”
Session 2022-24
Submitted by
S.Purna Chandra Rahul Rao
Roll No.: - 500107622059
Enrollment No.: - CC0595
MBA II Semester (Section- ‘C’)
2022-2024
DEPARTMENT OF MANAGEMENT
BHILAI INSTITUTE OF TECHNOLOGY
(Seth Balkrishan Memorial), (Accredited by NBA – AICTE)
(An ISO 9001:2000 Certified Institute), Bhilai House, Durg – 491001 (C.G.)
Appendix – I
Table of Contents
Sr. No. Particular Page. No.
1 Introduction to Multivariate Analysis 1-6
Mean Variance 2
Covariance 4
Correlation 5
Conjoint Analysis 15
Application of Conjoint Analysis in the field of Marketing 15
Discriminant Analysis 16
Bibliography 33
Appendix – II
List of Symbols
List of Abbreviations
Basic multivariate statistics involve the application of statistical methods to analyze and
interpret data with multiple variables. These methods allow researchers to explore
relationships, patterns, and dependencies among several variables simultaneously. Here are
some basic multivariate statistical techniques:
1
8. Multivariate Regression Analysis:
Extends simple linear regression to multiple independent variables.
9. Multidimensional Scaling (MDS):
Visualizes the similarity/dissimilarity between objects or cases in a dataset in a lower-
dimensional space.
These techniques are widely used in fields such as psychology, sociology, biology, finance,
and many others where researchers need to analyze data with multiple variables. The choice
of a specific technique depends on the nature of the data, the research question, and the
assumptions underlying the statistical method. Basic multivariate statistics are fundamental
for gaining insights into complex relationships within multidimensional datasets.
Mean Variance
"Mean variance" usually refers to two separate concepts in statistics and finance:
1. Mean:
The mean, often referred to as the average, is a measure of central tendency. It is
calculated by summing up all the values in a dataset and then dividing by the number
of observations. The mean provides a representative value around which the data
points tend to cluster.
Mathematically, the mean (x̄ ) is calculated as:
𝐧
̅ = ∑𝐢=𝟏 𝐗𝐢
𝐗
𝐧
Where 𝐱𝐢 represents each individual data point, and n is the total number of data
points.
2. Variance:
Variance is a measure of the dispersion or spread of a set of values. It quantifies how
far each data point in the dataset is from the mean. A high variance indicates that the
values are spread out over a wider range, while a low variance suggests that the values
are closely clustered around the mean.
Mathematically, the variance (s2 or σ2) is calculated as:
𝒏
∑𝒊=𝟏 (𝐱𝐢− 𝐱̅)
s2 = 𝒏
In finance, a related concept is often used called "portfolio variance," which involves
calculating the variance of a portfolio of assets.
2
3. Mean-Variance Portfolio Theory (in Finance):
In the context of finance, "mean-variance" is often associated with Harry Markowitz's
Modern Portfolio Theory (MPT). MPT suggests that an investor can construct a
portfolio of assets that optimizes the trade-off between expected return and risk
(measured by variance). The goal is to create a portfolio that maximizes expected
return for a given level of risk or minimizes risk for a given level of expected return.
In summary, "mean variance" can refer to the individual concepts of mean and variance, or it
can be associated with the optimization of investment portfolios in finance through Modern
Portfolio Theory.
Mean
Interpretation: After calculating the given dataset in MS Excel the Mean value of the given
dataset is 5.94
Mean
3
Interpretation: After calculating the given dataset in MS Excel the Mean value of the given
dataset is 31.84
Covariance
𝐧 ̅ ) (𝐘𝐢 −𝐘
̅)
∑𝐢=𝟏 (𝐗 𝐢 −𝐗
cov (X, Y) = 𝐧−𝟏
̅ and 𝐘
where 𝐗𝐢 and 𝐘𝐢 are individual data points, 𝐗 ̅ are the means of X and Y, and n is the
number of data points.
One limitation of covariance is that its value is not standardized, making it challenging to
interpret the strength of the relationship.
4
Positive covariance: Indicates that two variables tend to move in the same direction.
Negative covariance: Reveals that two variables tend to move in inverse directions.
Interpretation: After calculating the given dataset in MS Excel the Covariance of the given
dataset is 1.7
Another example:
Interpretation: After calculating the given dataset in MS Excel the Covariance of the given
dataset is 2.48
Correlation
Correlation provides a standardized measure of both the strength and direction of the
relationship, making it a more interpretable metric. Correlation is often preferred in practice,
especially when comparing relationships between variables on different scales.
5
Correlation is a standardized measure of the strength and direction of the linear relationship
between two variables. It scales the covariance by the product of the standard deviations of
the two variables. The correlation coefficient (r) is defined as follows:
𝒄𝒐𝒗 (𝑿,𝒀)
r=
𝝈𝒙𝝈𝒚
y: 8, 10, 12, 14
The problem can be fixed by dividing the covariance by the standard deviation to get the
correlation coefficient.
Interpretation: After calculating the given dataset in MS Excel the Correlation of the given
dataset is 0.98
6
the data. Linear combinations are prevalent in various statistical techniques, including
regression analysis, factor analysis, and linear algebra-based methods.
Here's a general form of a linear combination of variables for two variables, X and Y:
Linear Combination = aX + bY
In this expression:
1. Regression Analysis:
In simple linear regression, the predicted value (Y) is a linear combination of the
independent variable (X) and a constant term (𝒃𝟎):
Y = 𝒃𝟎 + 𝒃𝟏X
In multiple linear regression with multiple independent variables (𝑿𝟏, 𝑿𝟐, ..., 𝑿𝒏), the
predicted value (Y) is a linear combination of these variables and a constant term:
Y = 𝒃𝟎 + 𝒃𝟏𝑿𝟏 + 𝒃𝟐𝑿𝟐 + ... + 𝒃𝒏𝑿𝒏
2. Factor Analysis:
In factor analysis, observed variables are assumed to be linear combinations of
underlying latent factors and unique factors. The model expresses each observed
variable as a sum of its loadings on the latent factors:
𝐗𝐢 = 𝐚𝐢𝟏𝐅𝟏 + 𝐚𝐢𝟐𝐅𝟐 + ... + 𝐚𝐢𝐦𝐅𝐦 + 𝐄𝐢
where 𝐗𝐢 is the observed variable, 𝐅𝟏, 𝐅𝟐, ..., 𝐅𝐦 are latent factors, 𝐚𝐢𝟏, 𝐚𝐢𝟐, ..., 𝐚𝐢𝐦 are
factor loadings, and 𝐄𝐢 is the unique factor.
3. Linear Algebra:
In linear algebra, a system of linear equations can be represented as a matrix-vector
equation, which is essentially a set of linear combinations:
Ax = b
where A is a matrix, x is a vector of variables, and b is a vector of constants.
7
Linear combinations are fundamental in statistical modeling, providing a way to express
relationships between variables and factors in a linear fashion. They are often used to
simplify models, identify patterns, and make predictions based on the relationships observed
in the data.
8
UNIT – II
Logistic
Regression
Course Contents:-
Logistic Regression analysis and its application.
Logistic Regression
Logistic Regression is a statistical method used for binary classification problems, where the outcome
variable is categorical and has two classes (e.g., 0 or 1, Yes or No, True or False). Despite its name,
logistic regression is used for classification rather than regression.
In logistic regression, the logistic function (also called the sigmoid function) is used to model the
relationship between the independent variables and the probability of a particular outcome. The
logistic function is defined as:
𝟏
P(Y=1) =
𝟏+ 𝐞−(𝐛𝟎 + 𝐛𝟏𝐗𝟏 + 𝐛𝟐𝐗𝟐 + ...+ 𝐛𝐧𝐗𝐧)
Here:
The logistic regression model transforms the linear combination of independent variables using the
logistic function to ensure that the predicted probabilities fall between 0 and 1.
1. Medical Diagnosis:
Predicting whether a patient has a particular medical condition based on various
diagnostic measurements or test results.
2. Credit Scoring:
Assessing the likelihood of a customer defaulting on a loan or credit card payment
based on their credit history and financial information.
3. Marketing:
Predicting whether a customer is likely to purchase a product or respond to a
marketing campaign based on demographic and behavioral data.
9
4. Fraud Detection:
Identifying fraudulent activities, such as credit card fraud or online transaction fraud,
by analyzing patterns and anomalies in the data.
5. Employee Attrition:
Predicting whether an employee is likely to leave a company based on factors like job
satisfaction, tenure, and performance metrics.
6. Political Science:
Predicting election outcomes or political preferences based on polling data and
demographic information.
7. Customer Churn:
Predicting whether a customer is likely to churn (stop using a service) based on their
usage patterns and feedback.
8. Natural Language Processing (NLP):
In sentiment analysis, logistic regression can be used to classify text as positive or
negative.
Logistic regression is a widely used and versatile statistical method due to its simplicity,
interpretability, and effectiveness in a variety of applications where the outcome variable is
binary. It is important to note that logistic regression assumes a linear relationship between
the independent variables and the log-odds of the outcome.
• in a logistic regression model, dependent variable may have only two categories
which are binary and are coded as 0 and 1 (in case of a binary logit model).
• Data was collected from 26 respondents. Their purchase is dependent upon three
independent variables – attitude for shopping, age and income.
10
• We will have to build a logistic regression model where dependent variable intention
to purchase has two categories 0 and 1.
• All the independent variables are metric in nature. Logistic regression model has also
got the capacity of including some non metric independent variables.
• Note that logistic regression model graph is not linear in pattern. It is somewhat S-
shaped and obviously generates predictions that are always in between 0 and 1.
• In a logistic regression case, there will be only two possible outcomes: purchase and
no-purchase.
• So, in case of a logistic regression model, the probability of purchase will be modeled
as below:
11
Above figure exhibits Omnibus Tests of Model Coefficient for logit regression.
The model indicates whether all the variables have been included in the logit
regression equation or not.
Block indicates that all the variables have been included in the current block.
Model summary:
-2 Log likelihood value indicates how well the model fits the data. Smaller value of
this statistic is always desired. In fact, smaller the value of this statistic better the
model is.
The statistic is actually used to test the overall significance of the model.
The difference between -2 Log likelihood for a null hypothesis model (a model where
all the coefficient values are set as zero) and a best fit model is chi square distributed
with degrees of freedom equivalent to the number of independent variables included
in the model.
12
Classification Table:
• From the classification table, this is very clear that for purchase (indicated by 0),
12/13= 0.9230 = 92.30% of the cases are correctly classified.
• Like discriminant analysis, in logistic regression also we could have taken analysis
sample and validation sample.
shows variables in the equation for purchase problem. Like OLS regression, these are the
coefficient values of constant and other independent variables for estimating the dependent
13
variable with the help of independent variables. So, in case of purchase problem, the logistic
regression equation will take following form:
14
UNIT – III
Conjoint Analysis
In a conjoint analysis, respondents are presented with various product or service profiles that
differ in terms of attributes such as price, brand, features, and other relevant factors.
Respondents are then asked to rank or rate these profiles, providing insights into the relative
importance of each attribute and how they trade off against one another.
Applications in Marketing
1. Product Development:
Conjoint analysis helps businesses design new products by identifying the optimal
combination of features that will appeal to the target market. It provides insights into
which features are most valued by customers and how changes in one feature affect
overall product preference.
2. Pricing Strategy:
Businesses can use conjoint analysis to determine the optimal pricing strategy. By
assessing how price changes impact consumer preferences, companies can set prices
that maximize customer satisfaction and profitability.
3. Brand Positioning:
Conjoint analysis helps in understanding how different brand elements (such as brand
name, logo, or reputation) contribute to overall product preference. This information
aids in developing effective brand positioning strategies.
15
4. Market Segmentation:
Conjoint analysis can be used to identify distinct market segments based on consumer
preferences. This allows businesses to tailor their marketing strategies to specific
customer groups with similar preferences.
5. Feature Optimization:
For products with multiple features, conjoint analysis helps in optimizing the
combination of features to meet customer expectations. Businesses can identify which
features are must-haves and which are nice-to-haves.
6. Advertising and Communication Strategy:
Conjoint analysis assists in understanding how different advertising messages or
communication strategies influence consumer preferences. This information helps in
crafting more effective marketing messages.
7. Service Design:
In the service industry, conjoint analysis can be applied to understand customer
preferences for various service attributes, helping companies design services that
align with customer needs.
8. Market Entry and Expansion:
When entering new markets or expanding product lines, conjoint analysis aids in
tailoring offerings to the preferences of the target audience, ensuring a better chance
of success.
Conjoint analysis is a powerful tool for marketers to make informed decisions about product
and service offerings, allowing them to align their strategies with customer preferences and
improve overall market competitiveness.
Discriminant Analysis
16
There are two main types of discriminant analysis:
17
Steps in Discriminant Analysis:
1. Data Collection:
Gather data on the predictor variables for each observation and the corresponding
group to which each observation belongs.
2. Data Preprocessing:
Standardize or normalize the variables if necessary to ensure that they are on the same
scale.
3. Computing Discriminant Functions:
For LDA, compute linear discriminant functions that maximize the separation
between groups. For QDA, compute quadratic discriminant functions.
4. Classification:
Assign each observation to the group for which the computed discriminant functions
are maximized.
5. Assessment:
Evaluate the performance of the discriminant analysis using appropriate metrics, such
as accuracy, precision, recall, or the confusion matrix.
1. Finance:
Predicting the creditworthiness of individuals or companies based on financial
indicators.
2. Marketing:
Classifying customers into segments based on purchasing behavior or demographics.
3. Biology and Medicine:
Identifying the factors that discriminate between different disease groups or patient
outcomes.
4. Image and Speech Recognition:
Recognizing and classifying images or speech patterns.
5. Quality Control:
Identifying defects or quality issues in manufacturing processes.
18
6. Criminal Justice:
Predicting the likelihood of recidivism or classifying individuals into risk categories.
Discriminant analysis is a valuable tool when dealing with problems of classification and
group separation, and it has applications in various fields where distinguishing between
different groups is essential.
19
UNIT – IV
MANOVA and
Factor Analysis
Course Contents:-
Multivariate Analysis of Variance: Statistics and its application. Factor analysis: Exploratory
Factor Analysis (EFA) and Confirmatory Factor Analysis (CFA).
Statistics of MANOVA
20
Application of MANOVA:
1. Social Sciences:
MANOVA is often used in social sciences to analyze data where multiple dependent
variables are measured across different groups. For example, it could be used to
examine the impact of an educational intervention on students' performance in
multiple subjects.
2. Marketing Research:
In marketing, MANOVA can be applied to assess the effect of different advertising
strategies on various brand perception variables simultaneously.
3. Health Sciences:
MANOVA is used in health sciences to investigate the impact of different treatments
or interventions on multiple health outcomes or biomarkers.
4. Environmental Studies:
In environmental studies, MANOVA can be used to analyze the effects of various
factors (e.g., pollution levels, temperature) on multiple ecological variables.
5. Manufacturing and Quality Control:
MANOVA can be employed to assess whether the mean values of multiple quality
control measures differ across different manufacturing processes or locations.
6. Psychology and Education:
MANOVA is commonly used in psychology and education research to investigate the
impact of interventions or treatments on multiple dependent variables, such as
cognitive test scores or behavioral outcomes.
7. Biological Research:
MANOVA is used in biology to analyze data where multiple measurements (such as
gene expressions or physiological parameters) are taken under different experimental
conditions or treatments.
MANOVA is a powerful tool when researchers need to examine the effects of categorical
independent variables on multiple continuous dependent variables simultaneously. It provides
a comprehensive analysis of group differences across several variables, offering insights into
the overall pattern of effects.
21
Factor Analysis
Factor Analysis is a statistical technique used to identify and analyze underlying latent factors
that contribute to the observed correlations among a set of variables. It is commonly used in
data reduction and dimensionality reduction, aiming to explain the variance in a set of
observed variables using a smaller number of latent factors.
Steps in EFA:
1. Data Collection:
Collect data on a set of variables.
2. Correlation Matrix:
Examine the correlation matrix to understand the relationships among variables.
3. Factor Extraction:
Use statistical methods (e.g., principal component analysis or maximum likelihood) to
extract factors that account for the variance in the observed variables.
4. Factor Rotation:
Rotate the factors to make them more interpretable and easier to understand.
5. Factor Interpretation:
Interpret the meaning of the factors based on the variables that load strongly on each
factor.
EFA is exploratory in nature, allowing researchers to uncover patterns in the data without
preconceived notions about the number or nature of latent factors.
22
Steps in CFA:
1. Hypothesis Formulation:
Formulate a hypothesis about the structure of the latent factors based on theory or
previous research.
2. Model Specification:
Specify the relationships between the latent factors and the observed variables in a
measurement model.
3. Parameter Estimation:
Use statistical methods (e.g., maximum likelihood estimation) to estimate the
parameters of the model.
4. Model Evaluation:
Evaluate the fit of the model to the data using fit indices such as chi-square,
Comparative Fit Index (CFI), and Root Mean Square Error of Approximation
(RMSEA).
5. Model Modification:
If necessary, modify the model based on fit indices or theoretical considerations.
6. Final Model Interpretation:
Interpret the final model, considering the relationships between latent factors and
observed variables.
CFA is confirmatory, aiming to test a specific hypothesis about the factor structure rather
than exploring the data for potential factors.
23
Social Sciences:
Understanding latent factors contributing to attitudes, beliefs, or behaviors in social
research.
Finance:
Analyzing underlying factors influencing financial markets or investment decisions.
Factor Analysis is a versatile method widely used in various fields to uncover and understand
latent structures that may not be directly observable but explain the patterns in observed data.
24
25
26
27
28
UNIT – V
Structural
Equation
Modeling
Course Contents:-
Formulate Structural Equation Modeling using AMOS.
29
7. Path Model:
Specify the structural paths that represent the hypothesized relationships between
latent constructs. Define regression coefficients indicating the strength and direction
of the relationships.
8. Model Identification:
Ensure that the model is identified by providing sufficient information for the
estimation process. AMOS requires a fully identified model for accurate parameter
estimation.
9. Estimation Method:
Choose the estimation method, such as Maximum Likelihood (ML), for estimating
model parameters. AMOS will use this method to find the parameter values that
maximize the likelihood of observing the sample data.
10. Fit Indices:
Assess the goodness-of-fit using various fit indices, including chi-square,
Comparative Fit Index (CFI), Tucker-Lewis Index (TLI), Root Mean Square Error of
Approximation (RMSEA), and others. These indices provide information about how
well the model fits the observed data.
11. Model Modification:
If the initial model does not fit the data well, consider making modifications based on
modification indices provided by AMOS. These indices suggest improvements to the
model, such as adding or removing paths.
12. Interpretation:
Interpret the results of the SEM. Examine the standardized coefficients, factor
loadings, and significance levels to understand the strength and direction of
relationships. Consider theoretical implications and draw conclusions based on the
model fit.
13. Reporting:
Prepare a comprehensive report presenting the SEM results, including the model
diagram, parameter estimates, fit indices, and conclusions.
Remember that SEM in AMOS requires careful consideration of both the measurement and
structural components of the model. It is a iterative process where model modification might
30
be necessary to achieve a good fit. Additionally, a thorough understanding of the underlying
theory and careful model specification are crucial for the success of SEM analysis.
31
32
Bibliography
Books:
(1) Andy Field, “Discovering Statistics Using IBM SPSS Statistics”, Sage Publication
(2) Anil Kumar Mishra, “A Hand Book on SPSS for Research Work”, Himalaya Publishing House
(3) Lokesh Jasrai, “Data Analysis Using SPSS”, Sage Publications
Website:
(1) google.com
(2) ibm.com