0% found this document useful (0 votes)
9 views16 pages

Developing An Unsupervised Classification Algorithm For Characterization of Steel Properties (2012)

Uploaded by

sivipek239
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
9 views16 pages

Developing An Unsupervised Classification Algorithm For Characterization of Steel Properties (2012)

Uploaded by

sivipek239
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 16

The current issue and full text archive of this journal is available at

www.emeraldinsight.com/0265-671X.htm

IJQRM
29,4 Developing an unsupervised
classification algorithm
for characterization
368
of steel properties
Prasun Das
SQC and OR Division, Indian Statistical Institute, Kolkata, India, and
Shubhabrata Datta
Production Engineering Department,
Birla Institute of Technology Extension Centre, Deoghar, India

Abstract
Purpose – The purpose of this paper is to develop an unsupervised classification algorithm
including feature selection for industrial product classification with the basic philosophy of a
supervised Mahalanobis-Taguchi System (MTS).
Design/methodology/approach – Two novel unsupervised classification algorithms called
Unsupervised Mahalanobis Distance Classifier (UNMDC) are developed based on Mahalanobis’
distance for identifying “abnormals” as individuals (or, groups) including feature selection. The
identification of “abnormals” is based on the concept of threshold value in MTS and the distribution
property of Mahalanobis-D2.
Findings – The performance of this algorithm, in terms of its efficiency and effectiveness, has been
studied thoroughly for three different types of steel product on the basis of its composition and
processing parameters. Performance in future diagnosis on the basis of useful features by the new
scheme is found quite satisfactory.
Research limitations/implications – This new algorithm is able to identify the set of significant
features, which appears to be always a larger class than that of MTS. In industrial environment, this
algorithm can be implemented for continuous monitoring of “abnormal” situations along with the
general concept of screening “abnormals” either as individuals or as groups during sampling.
Originality/value – The concept of determining threshold for diagnostic purpose is algorithm
dependent and independent of the domain knowledge, hence much more flexible in large domain.
Multi-class separation and feature selection in case of detection of abnormals are the special merits of
this algorithm.
Keywords Unsupervised learning, Mahalanobis distance, MTS philosophy, Feature selection,
Threshold, Severity level, Genetic algorithms, Classification, Steel
Paper type Research paper

1. Introduction
In the process of classification system, by various parametric and non-parametric
methods, a classifier is designed to extract the distinguishing features and help for
International Journal of Quality decision making. A common approach in classification is to map the sparse
& Reliability Management high-dimensional attributes of objects into a dense low-dimensional space. Both types
Vol. 29 No. 4, 2012
pp. 368-383 of learning mechanisms, supervised and unsupervised, are applied on the given set of
q Emerald Group Publishing Limited information for the purpose of classification. In unsupervised learning, information on
0265-671X
DOI 10.1108/02656711211224839 input objects are used to fit a model without having any knowledge of output variable(s).
This method treats input objects as a set of random variables and builds a joint density An unsupervised
model for the data set. Another way of unsupervised learning is clustering of the input classification
patterns. Given a particular set of patterns or cost function, different clustering
algorithms lead to different clusters (Hinton and Sejnowski, 1999) such that each cluster algorithm
contains objects that share some important properties. In recent past, different
unsupervised classification methods are exemplified by various statistical concepts.
The most popular criteria include within group sums of squares underlying the single 369
stage method (Ward, 1963) and iterative relocation method called k-means clustering
(MacQueen, 1967). Other methods are hierarchical agglomeration based on the
classification likelihood (Murtagh and Raftery, 1984; Banfield and Raftery, 1992) and
expectation-maximization (EM) algorithm for maximum likelihood estimation of
multivariate mixture models (Celeux and Govaert, 1995).
In classification problems, whether supervised or unsupervised, identification of a
subset of important input features plays a significant role in creating clusters. The
selection of features is based on the application type and the kind of input data to be
used for classification. The usual way of doing so is to use some indices and evaluate
different feature subsets, which guides in selecting the best subset of features.
Recently, feature selection is becoming popular in unsupervised learning. To mention
some of the methods, sequential unsupervised feature selection algorithm, wrapper
approach based on EM algorithm, maximum entropy based method and the recently
developed neuro-fuzzy approach. Mitra et al. (2002) proposes an unsupervised
algorithm, which uses feature dependency/similarity for redundancy reduction, but
does not require any search. A new similarity measure, called maximal information
compression index, is used here in clustering.
Mahalanobis-Taguchi system (MTS), a kind of supervised technique, uses
Mahalanobis distance (MD) as a multivariate measure (Mahalanobis, 1936) for
prediction, diagnosis and pattern recognition in multi-dimensional system without any
assumption of statistical distribution (Taguchi and Jugulum, 2000, 2002), and attempts
to find out the significant features for generalization. The primary concept behind MTS
philosophy assumes only one group (cluster), namely “normal” and the corresponding
Mahalanobis space (MS) is obtained using the standardized variables of “normal” group
observations. This MS is used in identifying “abnormals”. Once this MS is established,
the number of features (variables) gets reduced using orthogonal array technique and
signal-to-noise (S/N) ratio by evaluating the contribution of individual feature. Recently,
some work has been done to compare MTS with neural network topologies
(Jugulum and Monplaisir, 2002; Hong et al., 2005; Das et al., 2006; Das and Datta,
2007). However, in all these works, the effectiveness of MTS to segregate “normal” and
“abnormal” observation seems to be heavily dependent on selecting the “normal” group
accurately. The determination of threshold value for future diagnosis as illustrated by
the developers is subjective and not clearly defined. Also, the method cannot prescribe
any standard procedure to distinguish the severity levels of abnormality.
In this work, two novel unsupervised classification algorithms called unsupervised
Mahalanobis distance classifier (UNMDC) are developed based on MD for identifying
“abnormals” as individuals (or, groups) including feature selection. The identification of
“abnormals” is based on the concept of threshold value in MTS and the distribution
property of Mahalanobis-D2. The performance of this algorithm is studied from various
perspectives on a real-life industrial environment, particularly for characterization
IJQRM of steel properties on the basis of its composition and processing parameters. The rest of
29,4 the paper is organized as follows. Section 2 contains the general discussion on MD and its
use in MTS. In Section 3, the proposed unsupervised classification scheme UNMDC is
presented. Section 4 narrates the implementation of the developed algorithm in real-life
steel processing system and demonstrates the results of classification including
selection of useful features. Some other merits of the proposed algorithm are highlighted
370 in Section 5. The paper is concluded in Section 6.

2. MD and its use in MTS


The MD was founded by a famous Indian statistician, P.C. Mahalanobis around
1936 (Mahalanobis, 1936). The MD is said to be a distance measure based on correlations
between the variables and by which different patterns could be identified and analyzed
with respect to a base or a reference point. MD is a very useful way of determining the
similarity of a set of values from an unknown sample to a set of values measured from a
collection of known samples. This measure is superior to Euclidean distance because it
takes distribution of the points (correlations) into account. Traditionally, it implies to
classify observations into different groups and is defined as follows:

D 2 tðxÞ ¼ ðx 2 mt ÞS 21
t ðx 2 m t Þ
t

where:
Dt is the generalized squared distance of each vector point from the t groups.
St represents the within-group covariance matrix.
mt is the vector of the means of the variables of the t groups.
X is the vector containing the values of the variables observed at point x.
When compared to other classical statistical approaches, the MD takes into account not
only the average value but also its variance and the covariance of the variables
measured. It accounts for ranges of the acceptability (variance) between variables.
It compensates for interactions (covariance) between variables. It is dimensionless.
If X
The distributional aspect of MD (D 2) is that it underlies Hotelling’s T 2 statistic. P
and S are the mean vector and covariance matrix of a sample of size n from Np [m, ],
then (n 2 1) ðX  2 mÞ , T 2( p, n 2 1), where Xp£ 1 , Np [m, S ] and
 2 mÞT S21 ðX
2
DX;  T 21 
 m ¼ ðX 2 mÞ S ðX 2 mÞ.
Taguchi was first to propose the use of MD (Taguchi and Jugulum, 2002) in a
multi-dimensional system for prediction, diagnosis and pattern recognition without any
assumption about the distribution, and attempted to find out useful variables for future
diagnosis, which has been referred to as MTS. In MTS, the MS (often called reference
group) is obtained using the standardized variables of “normal” data. The MS can be
used to discriminate between “normal” and “abnormal” data. Once this space is
established, the number of variables (or, attributes) is reduced, using orthogonal array
and S/N ratio, by evaluating the contribution of each attribute. Taguchi has also
incorporated the concept of a threshold for future diagnosis (Taguchi and Jugulum,
2000). The expected value of the MDi’s (MD for i-th pattern) for the “normal” items is
unity (Taguchi and Jugulum, 2002). This approximation is evidently based on
x2-distribution with p degrees of freedom. This is the probability distribution of pMDi,
provided the sampling is from multivariate normal distribution with known mean and An unsupervised
dispersion matrix. classification
algorithm
3. The proposed algorithms
3.1 Algorithm: UNMDC-I
This algorithm is developed in principle with MTS philosophy and the extension of
MTS principle, using statistical principles in an unsupervised manner, is discussed. 371
Assume that we have a (nxp) dataset where n is the number of observations and p is
the number of variables:
Stage-1. Construction of a measurement scale with MS as reference
1. Initialize K ¼ n; min. no. of patterns ¼ (p þ 1).
2. Initialize i ¼ 1.
3. While (i # K).
4. Take all K points except ith pattern from the dataset.
P
5. Find m and of the selected dataset.
P
6. Compute MDi for ith pattern based on m and .
7. Set i ¼ i þ 1; Go to Step 3.
8. Select group j if MDj ¼ max ðMDi Þ.
i
9. Set K ¼ K 2 1 and Dataset ¼ Previous dataset except the jth pattern.
10. If K . Min. no. of patterns, go to Step 2.
11. “normal” Group ¼ Dataset.
12. Compute Threshold value depending on “normal” group (Section 4.3 for details).
13. Create “normal” group with those patterns whose MDs are less than first
Threshold.
14. If number of points in the “normal” group changes then go to Step 12.
P
15. Find m and of the normal group.

Stage-2. Validation of measurement scale


16. Find MD for each pattern in the validation set.
17. Plot MD against the pattern number.
18. Detect abnormals of different severity by plotting threshold(s) values.

Stage-3. Identification of useful variables


19. Merge all the abnormals to form a group and use partial-F statistic to identify
those variables, which play important roles for separating abnormals from
“normal” group.

Stage-4. Future diagnosis with useful variable


20. For a test (new) pattern, compute its MD with respect to “normal” group.
21.Take decision, depending on the MD and threshold value, about the
characterization of the pattern either as “normal” or as “abnormal” with its
particular severity level.
IJQRM Note that this algorithm always converges to the same dataset, so it can be a good
algorithm for small dataset. Further, time complexity increases as the number of
29,4 observations becomes large. To overcome this, the second phase of this algorithm is
developed and discussed next.

3.2 Algorithm: UNMDC-II


372 .
Standardize the dataset either within [0, 1] or within [2 1, þ 1].
.
Divide the dataset into two parts: training and test set.
.
Set m ¼ Median of the set and:
X 1X n
¼ ðxi 2 mÞðxi 2 mÞ0
n i¼1
P
.
Compute MD for all points using m and .
.
Select first p þ 1 points corresponding to Min (MD).
P
.
Compute m and of these points.
P
.
Compute MD for all points using m and .
.
Remove those points whose MD . Cutoff.
.
Return the whole training dataset.
.
Set no. of cluster ¼ K; Iteration ¼ I; Epoch ¼ E; i ¼ 1.
.
while (i , ¼ I).
.
j ¼ 1; create K meanPvectors and K covariance matrix. Initial choice for
mi ¼ mean of the set; i ¼ Ip.
.
while ( j , ¼ E)
Select one point randomly from the dataset. Compute MD for that point from
each cluster.
.
Update ml if MDl ¼ min ðMDi Þ and store that point into the cluster of minimum
i
distance.
If training dataset is not empty then go to Step 14.
Update the covariance matrices of each group by corresponding cluster covariance.
j ¼ j þ 1; go to Step 13.
q ¼ 1;
. Take qth point from the test set. Compute the MD from each group.
Allocate it to group m if:
MDm ¼ min f MDi g and q ¼ q þ 1;
1#i#K
.
If q is not greater than the number of points of test set then go to Step 15.
.
Store the result in an array.
.
i ¼ i þ 1, go to step 11.
.
Take mode of that array (treated as convergence criteria) and predict the result of
the Test set.

The flowchart of the algorithm UNMDC-II is shown in Figure 1.


4. Implementation and results An unsupervised
The performance of the proposed algorithm vis-à-vis MTS methodology is evaluated classification
through different types of steel compositions, process parameters and the corresponding
properties of steel. Three sets of data are used for this purpose, each of which has got its algorithm
own specialities. The outcome of four stages of classification scheme (as discussed
in Section 3) is reported step-by-step in terms of average MDs for both the methods,
373
Start

Stardardise the Dataset

Yes
Outlines
Exclude Return Remaining
Return total Dataset No Outliers Dataset
D-Dataset
No of Cluster = K
predefined No = PN
Iteration No = IN
Initialize
INo = 1 and PNo = 1

mi = Mean of the Dataset


Σi = Identity Matrix ∀i = 1(1) K

Select One Point from the Dataset,


Compute MD ∀i = 1(1)K

Updata mi if MDi = min{MDi}

Dataset No
Empty

Yes

Updata Σi ∀i = 1(1) K
INo = INo+1, Dataset = D

No
INo > IN

1
2
Figure 1.
(continued) Flowchart of UNMDC-II
IJQRM
29,4
2
1

Yes

374 Select One Point from the Dataset,


Compute MD ∀i = 1(1)K

Allocate the point to Group i if


MDi = min{MDi}

No
Dataset
Empty

Yes

Store the Result in an array


PNo = PNo+1

No
PNo>PN INo = 1

Yes

Calculate Mode from


the Array

Return Classification
Result

END

Figure 1.
thresholds values for UNMDC, feature selection for both the methods and future An unsupervised
diagnosis with UNMDC only. classification
Dataset-I: related to steel hardness (using UNMDC-I) algorithm
The demand of high strength steels combined with adequate ductility, formability and
fracture toughness is mainly observed in the sectors like automobile, defense and naval
applications. The alloy chemistry (C, Mn, Si, Cu, Ti and B), cold deformation, cooling rate 375
(CRT), aging time and aging temperature have been taken as input parameters
(features), whereas hardness is designated as the output variable. The purpose is to
improve the hardenability which in turn is expected to give rise a dual phase
microstructure of steel depending on the finish rolling condition and CRT. The chemical
analyses are done in atomic spectrometer. The hardness testing has been carried out in
Vicker’s hardness testing machine. Information on acceptable level of hardness values
(“normal”) is known for 150 observations and the rest 92 patterns of input parameters
yield develop nonconforming (“abnormal”) hardness only (Das and Datta, 2007).
The data are analyzed using MTS and UNMDC methods, considering Stage I and
Stage II first. The results on average MD values for “normal” as well as for “abnormal”
are reported in Table I along with threshold value(s) obtained in UNMDC and the
number of patterns (observations) falling in between the threshold values.
As seen from Table I, all the 150 observations fall below the threshold value 6.6821
and identified as “normal”. Out of 92 abnormals, 56 of them falls within (6.6821,
0.2603 £ 1012), 21 lie between (0.2603, 7.4027) £ 1012, and 15 fall above 7.4027) £ 1012.
The overall average MDs of “normal” and “abnormal” are found as 0.9933 and
5.6262 £ 1012, both for MTS and UNMDC. The partition of “abnormal” group into three
sets (56, 21, 15) using UNMDC indicates different severity levels. Note that, the values
of threshold for MTS are domain knowledge dependent; therefore it is not possible to
report in Table I. The plot of scaled MD for both the methods is shown in Figure 2 over
242 set of points along with the horizontal lines indicating threshold values.
This plot indicates the level of separation of abnormal points from “normal” group
based on severity level. Looking at the plots of MDs (scaled) for both the methods, it is
observed that our proposed unsupervised algorithm maps the patterns quite
consistently with that of MTS in terms of scaled MD values.
We now show the results of selection of useful variables (features) as described in
Stage-III of both the methods for the steel data considered. The features are selected on
the basis of positive S/N ratio values in case of MTS and for UNMDC, the corresponding
high partial F-values indicate the importance of features to be selected (Table II).

Dataset-II: related to steel strength (using UNMDC-II)


The chemical composition and the amount of mechanical working mainly determine the
mechanical properties, particularly the strength of hot rolled plain carbon steel. In this
study, the chemical composition of the hot rolled plain carbon steel plate and the quality

Pattern type MTS (mean) UNMDC (mean) Threshold (UNMDC) No. of patterns Table I.
Average MDs for
Normal 0.9933 0.9933 6.6821 150 “normal” and “abnormal”
Abnormal 5.6262 £ 1012 5.6262 £ 1012 (0.0188, (0.2603, 7.4027) £ 1012 (56, 21, 15) patterns and thresholds
0.7403, 2.4588) £ 1013 (UNMDC)
IJQRM 1
29,4 0.9
UNMDC
0.8 MTS

0.7

376
Scaled MD
0.6

0.5

0.4

0.3

0.2

0.1
Figure 2.
Plot of scaled MD (MTS 0
0 50 100 150 200 250
and UNMDC) for Dataset-I
Patterns

MTS UNMDC
Selected Selected
features Gain (S/N ratio) features partial F-value

C, Mn, Si, Ti, B, Cu, Aging (213.4377, 213.0483, C, cold (5.2706, 0.1028, 0.9991,
CRT, cold temp. and 2 6.9996, 2 6.5296, deformation 0.6938, 1.4311, 2.1806,
Table II. deformation, aging aging time 2 5.6647, 2 6.8327, and aging time 0.0002, 38.2853,
Selection of useful temperature and 2 0.1486, 2 0.4957, 0.6679, 35.3996)
variables (features) aging time 4.8658, 942.7802)

of the finished plates based on their mechanical properties of an industrial situation are
considered. After the plates are manufactured, samples are taken for test piece analysis
with respect to their mechanical properties. Depending on the conformance to the
specifications of these sample plates, the corresponding heat for steel making is certified
as “OK” or “Diverted”. A total of 69 observations coming from test piece analysis of a
plate rolling mill contain information on carbon, manganese, sulphur, phosphorus and
silicon, as a result of ladle sample analysis along with the status of steel plates
(normal/“Ok” and abnormal/“Diverted”). No significant variation in the rolling
parameters is observed for this particular data set. Out of a total of 69 heats, 41 are from
“Ok” heats and 28 are declared as “Diverted” (Das et al., 2006).
The analysis of data using UNMDC-II yields the following result of classification
(Table III).
Partial-F statistic values for the five compositions C, Mn, S, P and Si are found as
9.0961, 62.5109, 0.5844, 2.1533 and 2.2699, respectively. The magnitude of these values
implies that Carbon (C) and Manganese (Mn) are the responsible (significant) features
for group separation (Figure 3).
The same dataset is used for comparison with other techniques. The results are
given in Table IV.
5. Merits of UNMDC scheme An unsupervised
We now discuss the advantages of the newly developed unsupervised algorithm, classification
particularly with respect to domain knowledge extraction using feature selection
methodology along with determining multiple threshold values representing different algorithm
severity levels of classification system from the “abnormal” group of measurements.
An algorithm for determining different threshold levels through this algorithm is also
described in this section. The practical implication of the merits of this unsupervised 377
classification algorithm in industrial environment is noteworthy.
5.1 Extraction of domain knowledge
5.1.1 Using UNMDC-I. To extract the domain knowledge based on the results of
feature selection by the algorithm, the input set of ten features of Dataset-I is applied to
this algorithm. The data set are classified into four groups (1: normal; 3: abnormals
with different severity/nonconforming levels) and for each group, the basic statistics
on hardness measurements are given in Table V.

Misclassification statistics
Ok ! Diverted Diverted ! Ok
Number Pattern no. Number Pattern no.
Table III.
3 10, 25, 41 1 61 Misclassification details
Overall 4 (6%) for steel plates

Mahalanobis Distance
1.2
Standardized MD

1
0.8 Normal Group
0.6 Abnormal Group

0.4
0.2
Figure 3.
0
Performance classification
1 6 11 16 21 26 31 36 41 46 51 56 61 66
based on MD values
Sample Points

Misclassification statistics
Ok ! Diverted Pattern no. Diverted ! Ok Pattern no. Total %

K-Means 3 22, 25, 41 6 46, 57,60, 61, 64, 65 9 3


Cart 3 29, 40, 41 1 61 4 6
LDF 2 25, 41 3 57, 61, 64 5 7
PamU 2 24, 41 6 46, 57, 60, 61, 64, 65 8 12
PamS 1 10 7 43, 51, 57, 61, 64, 65, 66 8 12
SOM 5 1, 10, 25, 33, 41 1 61 6 9 Table IV.
LVQ 3 22, 25, 41 6 46, 57, 60, 61, 64, 65 9 13 Misclassification details
UNMDC 3 10, 25, 41 1 61 4 6 for different techniques
IJQRM Interestingly, the UNMDC algorithm has been able to capture the information of
29,4 150 “normal” (acceptable) observations with respect to hardness measurements. To verify
the effectiveness of the group formation, Bartlett’s test for homogeneity of variances is
performed and no significant difference is found between the variances of different groups
with a value of test statistic equal to 3.132 ( p-value ¼ 0.372). Even, considering any
continuous distribution, Levene’s test is carried out yielding a value of test statistic equal
378 to 2.124 ( p-value ¼ 0.098), which confirms the equality of variances. When analysis of
variances is computed for hardness values, it shows that group means are significantly
different with F-value equal to 21.88 ( p-value ,0.0001). These findings with respect to
hardness property of steel also confirm the knowledge of materials science.
5.1.2 Using UNMDC-II. The same data (Dataset-I) is subjected to this algorithm in
an unsupervised way with the objective of classifying the information into three
groups, the algorithm generates the following information in Table VI.
To verify the effectiveness of the group formation, Bartlett’s test ensures the
homogeneity of variances among the groups with test statistic value of
3.152 ( p-value ¼ 0.207). Subsequently, through analysis of variances, it is seen that
the group hardness means are significantly different (F-value ¼ 25.94 with p-value
, 0.0001). These findings with respect to hardness property of steel also confirm the
knowledge of materials science.
Feature selection. The algorithm detects the first group as the “normal” group since
maximum points are allocated in that group. From the partial-F test it is evident that
Boron and CRT are the important features for separating first and second group where
as Manganese, Silicon, CRT, Cold deformation and Aging time are playing the
important role in separating first and third group. It is observed that CRT is most
important feature for group separation.

5.2 Threshold vs severity


5.2.1 Extension of MTS principle using UNMDC-I. It is observed that the process of
threshold determination for MTS is heavily domain knowledge dependent. The single
threshold limit used in MTS as discriminator for “normal” and “abnormal”
observations cannot be applied to identify different levels of abnormality. As we
have seen in our proposed UNMDC methodology, the first (lowest/minimum) threshold

Group No. of observations Average hardness SD

Normal 150 293.37 50.52


Table V. Abnormal(1) 56 240.36 56.56
Basic statistics for Abnormal(2) 21 222.86 53.39
hardness (UNMDC-I) Abnormal(3) 15 268.93 38.43

Group No. of observations Average hardness SD

Table VI. 1 121 249.14 56.79


Basic statistics for 2 73 295.38 47.75
hardness (UNMDC-II) 3 48 301.46 49.11
value isolates the “normal” group from abnormals whereas the next successive An unsupervised
threshold values determine severity levels among abnormals. classification
Higher is the threshold value, higher is the severity. No procedure has yet been
developed and exemplified to distinguish different levels of severity in abnormality. This algorithm
encourages in taking an attempt to develop an algorithm for determining different
threshold values which would represent severity levels. The algorithm is described below:
.
Sort MD values in ascending order. Let it have n values. 379
.
Compute Ratio (i ) ¼ ði Þ ¼ ðMDði þ 1Þ 2 MDði ÞÞ=MDði Þ ;i ¼ 1; . . . ; ðn 2 1Þ
and store it in the next column of MD.
.
Sort the ratio values in ascending order considering both the data-columns of
MD and ratio.
.
Cut ¼ x20:95;p =p where p is the dimension.
.
Initialize k ¼ 1, i ¼ n.
.
while (i . 0).
.
if MD(i ) $ Cut then Threshold (k) ¼ MD(i ) and Cut ¼ MD(i ).
.
k ¼ k þ 1, i ¼ i 2 1.
.
Go to Step 6.
.
Return Threshold.

5.2.2 Dataset-III: related to steel strength for multi-class classification. High strength low
alloy (HSLA) steels are usually low-carbon (up to 0.15 percent) steels with up to
1.5 percent manganese, low levels (under 0.035 percent) of phosphorus and sulphur
strengthened by small additions of elements, such as copper, nickel, niobium, nitrogen,
vanadium, chromium, molybdenum, silicon, or zirconium and sometimes by special
rolling and cooling techniques. The hot rolling parameters add to this complication and
also influence the final property. Proper adjustment of all these parameters can result
to an improved strength-ductility balance of such steels. HSLA steels are much
stronger and tougher than ordinary plain-carbon steels. The possibility of predicting
the strength of the steel after hot rolling would clearly be desirable and there have been
some attempts in this respect (Ichikawa et al., 1996; Ray Chaudhury and Das, 1997;
Datta and Banerjee, 2004; Das et al., 2006; Das and Datta, 2007).
To verify the performance of the UNMDC algorithm for determining multiple
threshold values, the dataset related to HSLA steel product is considered. This is
because the complexity of such steel system is more due to many input variables and
their interaction. Moreover, the dataset is relatively large as compared to plain carbon
steel system. So, the chance of falsification and the levels of severity could be more due
to unstable processing behaviour and inferior decision making while characterizing its
strength property. The dataset contains information of ten chemical compositions,
namely, Carbon (C), Manganese (Mn), Silicon (Si), Chromium (Cr), Nickel (Ni),
Molybdenum (Mo), Titanium (Ti), Niobium (Nb), Boron (B) and Copper (Cu) along with
six rolling parameters like slab reheating temperature (SRT), deformation in three
different temperature zones (D1, D2, D3), finish rolling temperature (FRT) and CRT.
Out of a total of 117 observations, 75 plates are from normal (high strength) group and
42 are from abnormal (low strength) group based on the strength property.
IJQRM Using the developed algorithm (Section 5.3.1) we get the following cutoffs (Table VII),
29,4 where the first one is from statistical point of view and others are data driven.
Accordingly, using the cutoff values against the computed MDs for all the
117 datapoint, Table VIII yields different severity levels, ranging from 2 to 6, for the
“abnormal” group.
The plot of MDs is shown in Figure 4 over 117 set of points along with the horizontal
380 lines indicating threshold (cutoff) values. This plot indicates the level of separation of
abnormal points from “normal” group based on severity level. From the following figure,
the existence of possible groups among abnormals is also observed.
Now, we explore the useful set of variables, considering each abnormal group (2-6) vs
normal group, based on S/N ratio analysis in the same way as done in Part-A work.

Cutoff , 1.9048 3.4979 14.5322 47.3934 86.3331 . 86.3331

Table VII. Severity 1 2 3 4 5 6


Multiple cutoff Total 69 22 8 7 3 8

Heat no. Severity Sl. no. Severity Heat no. Severity Heat no. Severity

1 1 31 1 61 1 91 3
2 1 32 1 62 1 92 2
3 2 33 1 63 1 93 2
4 1 34 1 64 1 94 3
5 1 35 1 65 1 95 4
6 1 36 1 66 1 96 4
7 1 37 1 67 1 97 3
8 2 38 1 68 1 98 3
9 1 39 1 69 1 99 3
10 1 40 1 70 1 100 4
11 1 41 1 71 1 101 3
12 1 42 1 72 1 102 4
13 1 43 2 73 2 103 3
14 1 44 1 74 1 104 6
15 1 45 1 75 1 105 5
16 1 46 1 76 2 106 5
17 1 47 1 77 2 107 5
18 1 48 1 78 1 108 6
19 2 49 1 79 2 109 4
20 1 50 1 80 2 110 4
21 1 51 2 81 2 111 6
22 1 52 1 82 2 112 6
23 1 53 1 83 2 113 6
24 1 54 1 84 2 114 6
25 1 55 1 85 2 115 6
26 1 56 1 86 2 116 6
27 1 57 1 87 2 117 4
Table VIII. 28 2 58 1 88 3
Severity/grade 29 1 59 1 89 2
identification of 117 heats 30 1 60 1 90 2
160 An unsupervised
140 3.5 classification
120 3 algorithm
100 2.5
MD

MD
80 2

60 1.5
381
40 1

20 0.5
Figure 4.
0 0 MD values and
0 20 40 60 80 100 120 10 20 30 40 50 60 70 80 90
different cutoff
Observations Observations

The purpose is to see the role of variables when groups of abnormals with different
severity exist. As before, based on the process knowledge, the eight variables (C, Mn, Si,
D1, D2, D3, SRT, FRT) are kept outside this analysis as they should be always present.
We have used the same L16 array where 16 trials were designed and for each trial, based
on the combination of “present” and “absent” level, MD values were calculated using the
data of each abnormal group. The result, in terms of gain for each (normal, abnormal
with a particular severity) combination, is shown in Table IX. The relevant points are
discussed in the next section.
Multi-class classification scheme using multiple cutoff search algorithm has been
quite successfully able to find the role of input variables for steel under study. It is clearly
revealed that Ni, Cu and post rolling CRT both individually and in combination have a
continuously increased gain with increase in the number of the group or with decreasing
the quality of the product. This suggests the fact these variables significantly control the
final property of the steel at all levels of its quality. On the other hand the carbide and
carbonitride precipitate formers like Cr, Mo, Ti, Nb and B have the effect only on the
higher quality of the product, as their gain is significant when compared with second
and third group only. Among them Cr and Mo is better influence converting the quality
of the steel from the third group to first and second groups, but Ti, Nb and B is important
for converting from the second group to first group. This finding also matches the
metallurgical understanding from the fact that Nb, Ti and B are better strengthener

Variables 2 vs 1 3 vs 1 4 vs 1 5 vs 1 6 vs 1

Ni 23.1876 39.0427 53.1281 56.7423 61.8316


Cu 11.2741 4.1075 21.7702 42.0785 60.0798
Ni-Cu 25.1893 24.7186 37.7166 45.8366 47.0745
CRT 7.7023 60.7162 72.6266 87.9983 86.6551
Ni-CRT 3.6225 34.6652 50.4983 53.9999 53.5233
Cu-CRT 6.1992 22.2712 41.8762 52.5912 49.2089
Cr 5.3357 26.0222 44.3726 40.6454 40.924
Mo 0.0124 10.9555 4.2181 0.2898 0.3584 Table IX.
Ti 5.8842 4.0005 3.2777 27.171 2 4.9727 Gain for different
B 14.1749 2.4395 2.7756 5.125 2 1.5568 variables in different
Nb 18.3902 1.7312 1.7031 27.1188 2 4.08 groups
IJQRM and only applied for achieving high and ultra-high strength steel. Here it is clearly
29,4 observed that the addition Nb, Ti and B has increased the strength of the steel to achieve
its maximum level.
The proposed multi-class classification scheme is able to split abnormals into
multiple groups with different severity levels based on the multiple threshold values.
The plot of MD values along with the cut-offs indicates the level of separation of
382 abnormal points from “normal” group. Performance in future diagnosis on the basis of
useful features (variables) is found quite satisfactory from the domain knowledge.
In the field of quality control, the decision making for the abnormal items might vary,
if they form different groups. This impacts the overall economy of the business from
the point of view of taking corrective actions. The technical reasons must also be
investigated for such different levels of abnormality.
6. Conclusion
The efficiency and effectiveness of the algorithm, as developed in this work, are observed
quite high when verified in a supervised manner. When compared to MTS philosophy,
the proposed algorithm satisfies its working principle in an unsupervised manner. It
shows consistency in selecting the patterns to form the so-called homogeneous group,
treated as “normal”. The algorithm uses MD as classifier on the basis of mean and
covariance structure of “normal” group. It is able to split “abnormals” into multiple
groups with different severity levels based on the multiple threshold values. The concept
of determining threshold for diagnostic purpose is algorithm dependent and independent
of the domain knowledge, hence much more flexible in large domain. This new algorithm
is able to identify the set of significant features, which appears to be always a larger class
than that of MTS. Performance in future diagnosis on the basis of useful features by the
new scheme is found quite satisfactory. In manufacturing set-up, this algorithm can be
implemented for continuous monitoring of “abnormal” situations along with the general
concept of screening “abnormals” either as individuals or as groups during sampling.
References
Banfield, J.D. and Raftery, A.E. (1992), “Ice floe identification in satellite images using
mathematical morphology and clustering about principles curves”, JASA, Vol. 87, pp. 7-16.
Celeux, G. and Govaert, G. (1995), “Gaussian parsimonious clustering models”, Pattern
Recognition, Vol. 28, pp. 781-93.
Das, P. and Datta, S. (2007), “Exploring the effects of chemical composition in hot rolled steel
product using MD scale under MTS”, Computational Materials Science, Vol. 38 No. 4,
pp. 671-7.
Das, P., Chaudhury, S., Bhattacharya, B.K. and Datta, S. (2006), “Efficiency of ANN over MTS in
classifying hot rolled steel product using small data”, Proceedings of International
Workshop on Neural Network and Genetic Algorithm in Material Science and Engineering,
BESU, Shibpur, India.
Datta, S. and Banerjee, M.K. (2004), “Kohonen network modeling for the strength of thermo
mechanically processed HSLA steel”, ISIJ International, Vol. 44 No. 5, pp. 846-51.
Hinton, G. and Sejnowski, T.J. (1999), Unsupervised Learning and Map Formation: Foundations
of Neural Computation, MIT Press, Cambridge, MA.
Hong, J., Taguchi, G., Paryani, K., Cudney, E.A., Jugulum, R. and Ragsdell, K.M. (2005),
“A comparison study of Mahalanobis-Taguchi system and neural network for
multivariate pattern recognition”, Proceedings of ASME IMECE 80029, Orlando, FL.
Ichikawa, K., Bhadeshia, H.K.D.H. and Mackay, D.J.C. (1996), “Model for solidification cracking An unsupervised
in low alloy steel alloy metals”, Science and Technology of Welding and Joining, Vol. 1 No. 1,
pp. 43-9. classification
Jugulum, R. and Monplaisir, L. (2002), “Comparison between Mahalanobis-Taguchi system and algorithm
artificial neural networks”, Journal of Quality Engineering Society, Vol. 10 No. 1, pp. 60-73.
MacQueen, J. (1967), “Some methods for classification and analysis of multivariate observations”,
Proceeding of the 5th Berkeley Symposium on Mathematical Statistics and Probability, 383
Vol. 1, University of California Press, Berkeley, CA.
Mahalanobis, P.C. (1936), “On the generalized distance in statistics”, Proceedings, National
Institute of Science of India, Vol. 2, pp. 49-55.
Mitra, P., Murthy, C.A. and Pal, S.K. (2002), “Unsupervised feature selection using feature
similarity”, IEEE Transactions on Pattern Analysis and Machine Intelligence, Vol. 24 No. 3,
pp. 301-12.
Murtagh, F. and Raftery, A.E. (1984), “Fitting straight lines to point patterns”, Pattern
Recognition, Vol. 17, pp. 107-20.
Ray Chaudhury, D. and Das, P. (1997), “Minimization of loss due to diversion of export grade
steel plates”, Quality Engineering, Vol. 10 No. 1, pp. 17-23.
Taguchi, G. and Jugulum, R. (2000), “New trends in multivariate diagnosis”, Sankhya: The Indian
Journal of Statistics, Vol. 62 No. 2, pp. 233-48 (Series B).
Taguchi, G. and Jugulum, R. (2002), The Mahalanobis-Taguchi Strategy: A Pattern Technology
System, Wiley, Hoboken, NJ.
Ward, J.H. (1963), “Hierarchical groupings to optimize an objective function”, JASA, Vol. 58,
pp. 234-44.

About the authors


Prasun Das is in the equivalent rank of Associate Professor in the SQC and OR Division of the
Indian Statistical Institute (ISI), Kolkata, India. He received his PhD (Materials Science and
Engineering) degree from BESU, Shibpur, India and Master’s degree in Technology with
specialties in quality, reliability and operations research from ISI. His responsibilities include
industrial consultancy, teaching and applied research in the field of quality control and
management, soft computing and mathematical programming. He is a certified lead assessor for
ISO 9000 QMS and engaged in various interdisciplinary research projects. He has published
many papers in national and international journals. Prasun Das is the corresponding author and
can be contacted at: [email protected]
Shubhabrata Datta is in the rank of Associate Professor in the Production Engineering
Department of Birla Institute of Technology Extension Centre, Deoghar, Jasidih, India. He
received his PhD from Bengal Engineering and Science University, Shibpur in 2003. His research
area includes application of soft computing techniques in the field of Materials Science and
Engineering. He has acted as guest editor in many international journal special issues and is a
member of the editorial board of an international journal.

To purchase reprints of this article please e-mail: [email protected]


Or visit our web site for further details: www.emeraldinsight.com/reprints

You might also like