Volume3 Issue12 PDF
Volume3 Issue12 PDF
Abstract: Automatic License Plate Recognition system is a real time embedded system which automatically recognizes the license
plate of vehicles. There are many applications ranging from complex security systems to common areas and from parking admission to
urban traffic control. Automatic license plate recognition (ALPR) has complex characteristics due to diverse effects such as of light
and speed. Most of the ALPR systems are built using proprietary tools like Matlab. This paper presents an alternative method of
implementing ALPR systems using Free Software including Python and the Open Computer Vision Library.
1. INTRODUCTION
The scientific world is deploying research in intelligent
transportation systems which have a significant impact on
peoples´ lives. Automatic License Plate Recognition (ALPR)
is a computer vision technology to extract the license number
of vehicles from images. It is an embedded system which has
numerous applications and challenges. Typical ALPR systems
are implemented using proprietary technologies and hence are
costly. This closed approach also prevents further research
and development of the system. With the rise of free and open
source technologies the computing world is lifted to new
heights. People from different communities interact in a Figure. 1 Example of a Number Plate with acceptable resolution
multi-cultural environment to develop solutions for mans
never ending problems. One of the notable contribution of the 2.2. Preprocess
open source community to the scientific world is Python.
Preprocessing is the set algorithms applied on the image to
Intel’s researches in Computer Vision bore the fruit called
enhance the quality. It is an important and common phase in
Open Computer Vision (OpenCV) library, which can support
any computer vision system. For the present system
computer vision development.
preprocessing involves two processes: Resize – The image
size from the camera might be large and can drive the system
2. PROPOSED SYSTEM
slow. It is to be resized to a feasible aspect ratio. Convert
Color Space – Images captured using IR or photographic
In India, basically, there are two kinds of license-plates, black
cameras will be either in raw format or encoded into some
characters in white plate and black characters in yellow plate.
multimedia standards. Normally, these images will be in RGB
The former for private vehicles and latter for commercial,
mode, with three channels (viz. red, green and blue).
public service vehicles. The system tries to address these two
categories of plates.[Reference 1]
2.1 Capture
The image of the vehicle is captured using a high resolution
photographic camera. A better choice is an Infrared (IR)
camera. The camera may be rolled and pitched with respect to
the license plates.
www.ijcat.com 756
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 756 - 761, 2014, ISSN:- 2319–8656
In image processing techniques, Haar-like features are used to 2.5. Optical Character Recognition
recognize objects from image . If our proposed system is
selected to detect only license plates then the Haar-like
features are used for this purpose and no further processing is Finally, the selected blobs are send to a Optical Character
done. This technique is old and laborious and more over needs Recognition (OCR) Engine, which returns the ASCII of the
a large database to store the collected samples nearly about license number.
10000 images of the plates and characters
3.WHY OPENCV??
2.3.2.2 License Plate Detection through Edge
Detection Advantages of OpenCV over MATLAB
In the other case, if our proposed system has to recognize
license plates, then the binary image is created from the Speed: Matlab is built on Java, and Java is built
image. After that following steps are performed to extract upon C. So when you run a Matlab program, your
license plate from binary image: computer is busy trying to interpret all that Matlab
1. Four Connected Points are searched from binary image. code. Then it turns it into Java, and then finally
executes the code. OpenCV, on the other hand, is
2. Width/Height ratio is matched against those connected basically a library of functions written in C/C++.
points. You are closer to directly provide machine
language code to the computer to get executed. So
3. License Plate region is extracted from image. ultimately you get more image processing done for
4. Transformation of extracted license plate is performed. your computers processing cycles, and not more
interpreting. As a result of this, programs written in
Then the extracted license plate is passed to next component OpenCV run much faster than similar programs
for further processing. written in Matlab. So, conclusion? OpenCV is damn
This approach is quick and takes less execution time and fast when it comes to speed of execution. For
memory with high a efficiency ratio. That’s why we have example, we might write a small program to detect
adopted this technique in our project peoples smiles in a sequence of video frames. In
Matlab, we would typically get 3-4 frames analyzed
per second. In OpenCV, we would get at least 30
frames per second, resulting in real-time detection.
www.ijcat.com 757
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 756 - 761, 2014, ISSN:- 2319–8656
www.ijcat.com 758
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 756 - 761, 2014, ISSN:- 2319–8656
5. The new algorithm proposed for character recognition because spell checks, valid word checks are
would give results with considerable percentage of errors on useless in case of number plates. so our
implementation. algorithm is simple, fast and occupies less
memory than an OCR engine. also it is
6. The efficiency of the proposed system can be measured
only in terms of number of license plates successfully and expected that it will provide correct results
correctly recognized which can only be measured upon upon implementation
implementation.
4.3 Proposed Algorithm
7. Efficiency and Performance of new system may decline due
to discard of OCR library but the memory requirements will
DESCRIPTION OF THE NEW ALGORITHM FOR
decrease and also the effort for installing, configuring and
running the system would decrease. CHARACTER RECOGONITION
l_mat[y][x]=g;
}
} 7. REFERENCES
www.ijcat.com 760
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 756 - 761, 2014, ISSN:- 2319–8656
www.ijcat.com 761
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
ABSTRACT
Background: dust phenomenon, especially in the summer, is a serious problem in Khuzestan province and has adverse effects on
health, environment and economic. Behavior change is the base for health associated risks prevention and one of the models for
behavior change in individual level is Health Belief Model. The aim of this study was to analyze the effect of educational
package on promotion of protective behaviors in exposure to dust phenomenon in Ahvaz teachers by SPSS software.
Methods: This was an experimental study in which 200 teachers randomly were divided into two groups, case and control groups
[n=100, in each group]
pha test. Before the educational intervention, questionnaire was
completed by two groups and educational requirements of subjects were detected and an educational package was designed and
implemented for 4 week. The control group received no intervention. After a month the effect of educational package on study
variables was evaluated. Data were analyzed with SPSS statistical software version 17, by descriptive and analytical tests.
Result: Mean age of case and control groups were 39.75±6.95 and 39.78±7.02 years, respectively. There was no significant
association between marriage and behavior, but there was a significant association between employment number of years and
behavior [p=0.03], education and behavior [p=0.03]. Based on the findings of this study there was a significant association
between the knowledge, health belief model components and behavior of the study subjects, before and after the intervention [p<
0.001].
Conclusion: designing and implementation of an educational package based on health belief model can promote the knowledge
and protective behaviors in exposure to dust particles effectively.
www.ijcat.com 762
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
The results showed that extract of these particles leaded to behaviors in exposure to dust phenomenon [11 questions,
DNA damage in these cells [6]. In a study on students of score range=11-55], respectively. In part 3, questions
850 schools in United Arab Emirates the prevalence of included: perceived susceptibility, perceived severity and
asthma and allergy were 13.6% and 73%, respectively and perceived benefits each with 7 questions [score range = 7-
there was a significant association between dust and 35], perceived barriers with 6 questions [score range = 6-
mentioned diseases [7]. Some recent epidemiological 36], cue to action with 2 questions [score range = 2-10] and
studies suggested that long-term transport of dust particles self-efficacy with 4 questions [score range = 4-20].
is associated with increased daily mortality in Seoul [8] and
Taipei and Taiwan [9] and respiratory and cardiovascular The validity of questionnaire was evaluated by means of
diseases [8]. although , the dust particles contribute to face validity and content validity methods. Face validity
drinking water contamination and therefore gastrointestinal was evaluated by means of relevance, simplicity and clarity
disease [10]. of questions. Questionnaire was evaluated by 10 experts
[included 5 experts in health education, 5 experts in
Dust occurrence increases some heavy metals such as lead environmental health, 1 epidemiologist and 2 experts in
concentrations up to 3-fold [11]. Also the concentration of Biostatistics]
toxic metals, mercury and arsenic, greatly increased [12]. Q
Air -borne microorganisms concentration in dusty days
increase and most of these microorganisms are pathogen that were same with study population in demographic
and cause disease. characteristics.
Iran is located in an area with a dry climate and over 30%
of the country is arid and semi-arid area [14]. In the past detected: Knowledge: 0.76, perceived susceptibility: 0.73,
few years the country has been exposed to the dust perceived severity: 0.88, perceived benefits: 0.72,
phenomenon. This country because of neighboring with the perceived barriers: 0.77, cue to action: 0.71, perceived Self-
wide expanse of desert is adversely affected by this efficacy: 0.71 and protective behaviors: 0.71.
phenomenon. One of the areas that has been affected by
this phenomenon is Khuzestan province that is located in The questionnaire was used before and after the educational
southwestern of Iran [15]. This phenomenon has been package implementation to determine the perceived
associated with some problems and rise in adverse effects knowledge, sensitivity, severity, benefits, barriers and self-
on health, environment and economic [15]. In the dusty efficacy, and behavior of subjects. Data were collected by a
days, admit of patients with pulmonary disease to health questionnaire in interview method before intervention in
centers in Ahvaz has showed 70 percent increase [1]; One case and control group. Then data were analyzed and the
way for reducing the incidence of diseases caused by dust educational needs of subjects were detected and
is educational interventions. Health education experts use educational package was designed. Educational package
the appropriate models in order to health education included an educational booklet, pamphlet and CD that
interventions design, one of this models is Health Belief represented essential information in relation to dust
Model. The aim of this study was design and particles, disease prevention and protective behaviors. Then
implementation of educational package based on health researcher represent educational package to the case group
belief model and evaluation of its effects on protective in four sessions [each sessions was 90 minutes].
behaviors in teachers by SPSS software. In this study Educational methods were the lecture, questioning and
educational package was an educational program that has responding and showing the video clip. Immediately and
been designed based on educational needs of subjects in two months after the educational intervention, subjects data
order to prepare the subjects for implementation of were collected by questionnaire and were analyzed. The
protective behaviors in exposure to dust phenomenon. control group received no intervention. Data were analyzed
with SPSS statistical software version 17, by frequency
distribution, correlation coefficient, t-Student, Chi-square,
Mann-Whitney and Repeated measures tests.
2. METHODS
This was an experimental and analytical study [before and
after] that has been conducted in Ahvaz. Two hundred 3. RESULTS
teachers randomly were divided into two groups, case and Two hundred teachers were participated in this study. Mean
control groups [n=100, in each group]. The inclusion age of case and control groups were 39.75±6.95 and
criteria included: employment for at least three years, lack 39.78±7.02 years; respectively. Age 40-49 years had the
of respiratory disease and cardiovascular disease and most frequency in the case group [46%] and the control
satisfaction for participation in the study. Exclusion criteria group [45%]. In the two groups more present of subjects
included: unsatisfaction for participation in the study and were married [82% in the case group and 81% in the
nonparticipation in the educational sessions. Data were control group]. In the two groups most of subjects had
collected by a questionnaire that was designed according to Bachelor's degree. More present of subjects had two
the health belief model constructs. The questionnaire children [47% in the case group and 46.3% in the control
contained 78 questions in four parts. This parts included group] and less present of subjects had four children. Most
questions regarding to individual characteristics [19 j ’
questions], knowledge [14 questions, score range=14-32], and protective behaviors. Age, marriage, education,
health belief model constructs [34 questions] and protective
www.ijcat.com 763
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
Table 1. Information sources regarding the dust Before the intervention, protective behaviors of teachers in
phenomenon in teachers, Ahvaz exposure to dust phenomenon were in intermediate level.
But significant difference between behavior score of cases
and controls after intervention showed the positive effect of
The mean of knowledge, perceived susceptibility, educational package on promotion of protective behaviors
in case group. In Araban et al. study after the intervention
Groups cases controls behavior score was significantly different between case and
control groups [20]. The results of Giles et al. meeting in
Canada on strategies for reducing the adverse effects of air
Information yes no yes no p-value pollution on health, entitled "The decision to effective
source intervention", showed that personal behavior modification
and pollutants exposure reduction are appropriate
Radio & 93 7 95 5 0.552 approaches for reducing the adverse effects of air pollution
television [21]. Sexton study showed that on dusty days persons
Newspaper & 43 57 40 60 0.667 changed their behavior by reducing time spent outdoors by
magazine 18% or 21 minutes [22].
I g ’
family 58 42 58 42 1
sources of information about protective behaviors in the
coworkers 57 43 58 42 0.886 exposure to dust particles were radio, television and family.
Significant difference between knowledge score of two
friends 55 45 57 43 0.776 groups after the educational intervention was due to the
Book & booklet 32 68 31 69 0.115 educational sessions about protective behaviors in exposure
to dust phenomenon and this educational sessions promoted
Physician and the knowledge of case group about protective behaviors.
staff of health These results are in line with the use of Health Belief
center 36 64 35 65 0.077
Model in researches about diabetes control and self-care
and promotion of knowledge after the educational
internet 47 53 45 55 0.777 intervention [23, 24]. Boonkuson et al. showed that
perceived severity, perceived benefits, perceived barriers, protective behaviors in exposure to health problems
perceived self-efficacy, cue to action and behavior score depends on the knowledge and attitude [25]. Pazira et al.
were not significantly different between cases and controls reported that a part of Tehran population knowledge about
before intervention. Whereas, immediately and two months air pollution and protective behaviors was in low level [26].
after the educational intervention there was a significant In the health belief model constructs the perceived
difference between cases and controls in mentioned susceptibility score before intervention was the same in
variables [p=0.001] [table 2, 3]. both groups. After intervention perceived susceptibility
, 16% ’ score was significantly different between case and control
home in dusty days but after the intervention 57% of cases groups [p=0.001]. This finding is consistent with increased
’ perceived susceptibility in researches about the
intervention 70% of cases sometimes educated their osteoporosis prevention [27] and diet care [24].
students in relation to air pollution but after the intervention Also, perceived severity score before intervention in two
75% of cases often educated their students. Before the groups sho ’
intervention, only 2% of cases have been eaten more of illnesses caused by dust particles was over the average,
amount of fruit and vegetable in dusty days but after the probably due to the illness of friends or coworkers or
intervention the rate increased to 41%. Before the damages caused by dust particles. The dramatic increase in
intervention, only 3% of cases have been eaten more the perceived severity score of the case group seems to be
amount of milk in dusty days. ’
www.ijcat.com 764
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
and providing educational package included showing video increased similarly [23, 27]. Also in the Praphant et al.
clip, booklet and pamphlet, mention to importance of study perceived severity was in moderate level [60.6%]
protective behaviors on the dusty days, high cost of [28].
pulmonary, cardiovascular and gastrointestinal tract
diseases. In the other studies perceived severity has been
Table 2. Comparing knowledge and behavior scores regarding the protective behaviors in exposure to dust phenomenon in
teachers, Ahvaz
Table 3. Comparing health belief model constructs scores regarding the protective behaviors in exposure to dust phenomenon in
teachers, Ahvaz
www.ijcat.com 765
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
sample t test
www.ijcat.com 766
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
4. Griffin DW, Kellogg CA. Dust storms and their impact 18. Taheri Aziz M. Effectiveness of Designed Health
on ocean and human health: E ’ Education Package on Healthy Behaviors of Patients with
EcoHealth. 2004;1[3]:284-95. Tuberculosis at Pasteur Institute of Iran [dissertation].
Tehran: Tarbiat modares of Medical sciences; 2004. p.67-
5. Peters A. Particulate matter and heart disease: evidence 8.[In persian]
from epidemiological studies. Toxicology and applied
pharmacology. 2005;207[2]:477-82. 19. Al-Hurban AE, Al-Ostad AN. Textural characteristics
of dust fallout and potential effect on public health in
6. Meng Z, Zhang Q. Damage effects of dust storm PM< Kuwait City and suburbs. Environmental Earth Sciences.
sub> 2.5</sub> on DNA in alveolar macrophages and lung 2010;60[1]:169-81.
cells of rats. Food and chemical toxicology.
2007;45[8]:1368-74. 20. Araban M. Design and Evaluation of a Theory-Based
Educational Intervention on Behavioral Improvement in
7. Bener A, Abdulrazzaq Y, Al-Mutawwa J, Debuse P. Pregnant Women in Terms of Exposure to Air Pollution
Genetic and environmental factors associated with asthma. [Dissertation]. Tehran: Tarbiat Modares University, Faculty
Human biology. 1996:405-14. of Medical Sciences; 2013. [Text in Persian]
8. Kwon H-J, Cho S-H, Chun Y, Lagarde F, Pershagen G. 21. Giles LV, Barn P, Kunzli N, Romieu I, Mittleman MA,
Effects of the Asian dust events on daily mortality in Seoul, van Eeden S, et al. From good intentions to proven
Korea. Environmental Research.2002;90[1]:1-5. interventions: effectiveness of actions to reduce the health
impacts of air pollution. Environmental health perspectives.
9. Ichinose T, Yoshida S, Hiyoshi K, Sadakane K, Takano
H, Nishikawa M, et al. The effects of microbial materials 2011; 119[1]:29.
adhered to Asian sand dust on allergic lung inflammation. 22. Sexton AL. Responses to Air Quality Alerts: Do
Archives of environmental contamination and toxicology. Americans Spend Less Time Outdoors? [Dissertation].
2008; 55[3]: 348-57. Minnesota: Department of Applied Economics, University
10. Kellogg CA, Griffin DW, Garrison VH, Peak KK, of Minnesota; 2011.
Royall N, Smith RR, et al. Characterization of aerosolized 23. Mohebi S, Sharifirad G, Hazaveyee S. The effect of
bacteria and fungi from desert dust events in Mali, West educational program based on Health Belief Model on
Africa. Aerobiologia. 2004;20[2]:99-110. diabetic foot care. Int J Diab Dev Ctries. 2007; 27:18-23.[In
11. Viana M, Kuhlbusch T, Querol X, Alastuey A, Harrison persian]
R, Hopke P, et al. Source apportionment of particulate
www.ijcat.com 767
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 762 - 768, 2014, ISSN:- 2319–8656
www.ijcat.com 768
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 769 - 773, 2014, ISSN:- 2319–8656
Abstract: In this paper, we propose an illumination-robust face recognition system using local directional pattern images. Usually,
local pattern descriptors including local binary pattern and local directional pattern have been used in the field of the face recognition
and facial expression recognition, since local pattern descriptors have important properties to be robust against the illumination
changes and computational simplicity. Thus, this paper represents the face recognition approach that employs the local directional
pattern descriptor and two-dimensional principal analysis algorithms to achieve enhanced recognition accuracy. In particular, we
propose a novel methodology that utilizes the transformed image obtained from local directional pattern descriptor as the direct input
image of two-dimensional principal analysis algorithms, unlike that most of previous works employed the local pattern descriptors to
acquire the histogram features. The performance evaluation of proposed system was performed using well-known approaches such as
principal component analysis and Gabor-wavelets based on local binary pattern, and publicly available databases including the Yale B
database and the CMU-PIE database were employed.
www.ijcat.com 769
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 769 - 773, 2014, ISSN:- 2319–8656
most of previous works were used the local pattern descriptors 2.2 2-D Principal Component Analysis
to extract the histogram features. However, we utilize the Principal component analysis is a well-known feature
transformed image from local pattern descriptor, i.e. LDP extraction and data representation technique widely used in
image as input image for further feature extraction procedure, the areas of pattern recognition, computer vision, signal
i.e. 2D -PCA algorithm. The advantage of the proposed processing, and so on. The central underlying concept is to
approach is that the illumination effects on face can be reduce the dimensionality of a data set while retaining the
degraded by using binary pattern descriptor, and also 2D-PCA variations in the data set as much as possible. In the PCA-
is more robust against illumination variation than global based face recognition method, 2D face image matrices must
features such as PCA and LDA since 2D -PCA is a line-based be previously transformed into 1D image vectors column by
local feature. In fact, we will be show that the recognition column or row by row fashions. However, concatenating 2D
accuracy of the proposed system outperforms that of matrices into 1D vector often leads to a high-dimensional
conventional approaches in the experimental results vector space, where it is difficult to evaluate the covariance
matrix accurately due to its large size. Furthermore,
2.1 Local Directional Pattern computing the eigenvectors of a large covariance matrix is
The LBP operator labels the pixels of an image by very time-consuming.
thresholding a (3x3) neighborhood of each pixel with the
center value and considering the results as a binary number, of To overcome these problems, a new technique called 2D-
which the corresponding decimal number is used for labeling. PCA was proposed, which directly computes eigenvectors of
The derived binary numbers are called local binary patterns or the so-called image covariance matrix without matrix-to-
LBP codes. While the LBP operator uses the information of vector conversion. Because the size of the image covariance
intensity changes around pixels, LDP operator use the edge matrix is equal to the width of images, which is quite small
response values of neighborhood pixels and encode the image compared with the size of a covariance matrix in PCA, 2D-
texture. The LDP is computed as follow. The LDP assigns an 8 PCA evaluates the image covariance matrix more accurately
bit binary code to each pixel of an input image. This pattern is and computes the corresponding eigenvectors more efficiently
then calculated by comparing the relative edge response values than PCA. It was reported that the recognition accuracy of
of a pixel by using Kirsch edge detector. Given a central pixel 2D-PCA on several face databases was higher than that of
in the image, the eight-directional edge response values mi (i = PCA, and the feature extraction method of 2D-PCA is
0, 1, ….,7) are computed by Kirsch masks as shown in Figure computationally more efficient than PCA. Unlike PCA, which
1. Since the presence of a corner or an edge shows high treats 2D images as 1D image vectors, 2D-PCA views an
response values in some particular directions, thus, most image as a matrix. Consider an m by n image matrix A.
prominent directions of k number with high response values ( n d )
are selected to generate the LDP code. In other words, top-k Let R be a matrix with orthonormal columns, n d.
directional bit responses, bi , are set to 1, and the remaining (8 Projecting A onto X yields m by d matrix Y= AX. In 2D-
- k) bits are set to 0. Finally, the LDP code is derived by PCA, the total scatter of the projected samples is used to
determine a good projection matrix X . Suppose that there are
M training face images, denoted m by n matrices Ak (k =1,
7
1 , x 0 A 1/ M Ak
b (m i m k ) 2 , bi ( x )
i
LDP k i
x 0, 2, …,M), and the average image is denoted as k
.
i0 0, (1) Then, the image covariance matrix, G is given by
m3 m2 m1 b3 b2 b1
It has been proven that the optimal value for the projection
matrix X opt is composed by the orthonormal eigenvectors
m4 X m0 b4 X b0 X1, X2 , …, Xd of G corresponding to the d largest
eigenvalues, i.e., Xopt = [ X1, X2 ,…, Xd]. Since the size of
m5 m6 m7 b5 b6 b7 G is only n by n, computing its eigenvectors is very efficient.
The optimal projection vectors of 2D-PCA, X1, X2,…,Xd are
used for feature extraction. For a given face image A, the
Figure 1. Edge Response and LDP Binary Bit Positions feature vector Y = [Y1, Y2, …,Yd] , in which Y has a
dimension of m by d , is obtained by projecting the images
into the eigenvectors as follows:
85 32 26 313 97 503 0 0 1
www.ijcat.com 770
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 769 - 773, 2014, ISSN:- 2319–8656
m d
) ( f i, j f i, j )
k k 2
d (F , F
i 1 j 1
(4)
where k is 1,2,…M , and M is the total number of training 3.1 Yale B Database
images. This distance measurement between 2D-PCA features
is further employed to classify unknown user. To evaluate the performance of the proposed method, we
partitioned the Yale B database into training and testing sets.
3. EXPERIMENTAL RESULTS Each training set comprised of seven images per subject, and
To evaluate the robustness of the proposed method, the remaining images were used to test the proposed method.
we used images from the Yale B database and CMU-PIE We selected the illumination-invariant images for training, and
database. In the Yale B database, we employ 2,414 face images the remaining images with varying illumination were
for 38 subjects representing 64 illumination conditions under employed for testing. Next, we investigated the recognition
the frontal pose, in which subjects comprised 10 individuals in performance of proposed approach with conventional
the original Yale face database B and 28 individuals in the recognition algorithms such as PCA and Gabor-wavelet based
extended Yale B database. The CMU-PIE database contains on LBP. For the Yale B database, the recognition results in
more than 40,000 facial images of 68 individuals, 21 terms of different pre-processing images and algorithms are
illumination conditions, 22 light conditions, 13 poses and four shown in Figure 4. To further disclose the relationship between
different expressions. Among them, we selected each the recognition rate and dimensions of feature vectors, we
illumination and light images of 68 individuals with frontal showed the recognition results along with different dimensions
pose (c27). So, the CMU-PIE illumination set consists of 21 in Figure 4. Also, we summarized the maximum recognition
images of 68 individuals (21x68 images in total), and the rates as various approaches in Table 1. As a result, the
CMU-PIE illumination set also consists of 22 images of 68 proposed approach using LDP and 2D-PCA shows a maximum
individuals (22x68 images in total). All face images of two recognition rate of 96.43%, when k is 3. However, the
databases were converted as grayscale and were cropped and maximum recognition rates revealed 81.34% and 69.50% for
normalized to a resolution of (48x42) pixels. Figure 3 show an PCA and Gabor-wavelets based on LBP approaches,
example of raw, histogram equalization, LBP, and LDP images respectively. Consequently, the recognition accuracy of
in CMU-PIE illumination database, respectively. Remark that proposed method was better than that of conventional methods,
LDP images are divided into different groups as k number. The and it also shows performance improvement ranging from
performance evaluation was carried out using each database of 15.09% to 29.63% in comparison to conventional methods.
the Yale B database and CMU-PIE illumination/light database
with each pre-processing images.
www.ijcat.com 771
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 769 - 773, 2014, ISSN:- 2319–8656
LDP
LDP (K=5) 81.34% 95.49% X (K=1) 84.71% 99.93% X
www.ijcat.com 772
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 769 - 773, 2014, ISSN:- 2319–8656
www.ijcat.com 773
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014, ISSN:- 2319–8656
Abstract- Grid computing provides a framework and deployment environment that enables resource
sharing, accessing, aggregation and management. It allows resource and coordinated use of various
resources in dynamic, distributed virtual organization. The grid scheduling is responsible for resource
discovery, resource selection and job assignment over a decentralized heterogeneous system. In the
existing system, primary-backup approach is used for fault tolerance in a single environment. In this
approach, each task has a primary copy and backup copy on two different processors. For dependent
tasks, precedence constraint among tasks must be considered when scheduling backup copies and
overloading backups. Then, two algorithms have been developed to schedule backups of dependent and
independent tasks. The proposed work is to manage the resource failures in grid job scheduling. In this
method, data source and resource are integrated from different geographical environment. Fault-
tolerant scheduling with primary backup approach is used to handle job failures in grid environment.
Impact of communication protocols is considered. Communication protocols such as Transmission
Control Protocol (TCP), User Datagram Protocol (UDP) which are used to distribute the message of
each task to grid resources.
1. INTRODUCTION
1.2 Grid Computing Overview
1.1 Grid Computing
A distributed heterogeneous
Grid Computing is distributed; large computing system consists of a distributed
scale cluster grid computing has emerged as suite of different high-performance machines,
the next-generation parallel and distributed interconnected by the high-speed networks, to
computing methodology, which aggregates perform different computationally intensive
dispersed heterogeneous resources for solving applications that have various computational
various kinds of large-scale parallel requirements. Heterogeneous computing
applications in science, engineering and systems range from diverse elements or
commerce. It can integrate and utilize paradigms within a single computer to a cluster
heterogeneous computational resources from of different types of personal computers to
different networks or regional areas into a high coordinate geographically distributed machines
performance computational platform and can with different architectures. Job scheduling is
solve complex computing-intensive problems one the major difficult tasks in a computational
efficiently. Grid service represents grid.
convergence between high performance
computation and web service. Grid aims
ultimately to turn the global network of
computers into a vast computational resource.
www.ijcat.com 774
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014, ISSN:- 2319–8656
www.ijcat.com 775
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014, ISSN:- 2319–8656
www.ijcat.com 776
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014, ISSN:- 2319–8656
Author:
Mrs.R.Kohila received
M.E(CSE) degree from
Kongu Engineering
College(Affiliated to
Anna University,
Autonomous),
Perundurai, India in 2011 and MCA
degree from Bannari Amman Institute
of Technology (Affiliated to Anna
University), Sathyamangalam, India,
in 2009 and B.Sc., degree from Trinity
College of Arts and Science for
women (Affiliated to Periyar
University), Namakkal,India, in 2006.
She has the experience in Teaching of
3+Years. Now she is currently
working as an Assistant Professor in
V.S.B Engineering College, Karur,
Tamil Nadu, and India. His research
interests include Data Mining,
Advanced Data Bases, Computer
Networks etc. She had presented
papers in 2 National Conferences so
far.
www.ijcat.com 777
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 778 - 781, 2014, ISSN:- 2319–8656
Abstract: The data centres and energy consumption characteristics of the various machines are often noted with different capacities.
The public cloud workloads of different priorities and performance requirements of various applications when analysed we had noted
some invariant reports about cloud. The Cloud data centres become capable of sensing an opportunity to present a different program.
In out proposed work, we are using a hybrid method for resource provisioning in data centres. This method is used to allocate the
resources at the working conditions and also for the energy stored in the power consumptions. Proposed method is used to allocate the
process behind the cloud storage.
Keywords: Cloud workload, Hybrid resource provisioning, Cloud storage and Invariant reports.
www.ijcat.com 778
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014
characteristics of this area which make the cloud computing algorithm is used to attempt each node keep busy and goal of
being cloud computing and distinguish it from other research load balance. Also proposed LBMM (Load Balance Min-
area. The cloud computing has its own technical, economic, Min) scheduling algorithm can make the minimum execution
user experience characteristics. The service oriented, loose time of each task on cloud computing environment and this
coupling, strong fault tolerant, business model and ease use will improve the load unbalance of the Min-Min. In order to
are main characteristics of cloud computing. Abstraction and reach load balance and decrease execution time for each node
accessibility are two keys to achieve the service oriented in the three-level cloud computing network, the OLB and
conception. In loose coupling cloud computing run in a client- LBMM scheduling algorithm are integrated. The load
server model. The client or cloud users connect loosely with balancing of three-level cloud computing network is utilized
server or cloud providers. Strong fault tolerant stand for main all calculating result could be integrated first by the
technical characteristics. The ease use user experience secondlevel node [5]
characteristic helps cloud computing being widely accepted
by non computer experts. These characteristics expose the In January 31, 2011, Sivadon Chaisiri, Bu-Sung Lee, and
essential of cloud computing. [1] Dusit Niyato discuss about the Optimization of Resource
Provisioning Cost. Under the resource provisioning optimal
In 2010 Pushpendra kumar pateria, Neha Marria discussed cloud provisioning algorithm illustrates virtual machine
resource provisioning in sky environment. Resource manager management that consider multiple provisioning stages with
is used for resource provisioning and allocate of resource as demand price uncertainty. In this task system model of cloud
user request. Offer the rule based resource manager in sky computing environment has been thoroughly explained using
environment for utilization the private cloud resource and various techniques such as cloud consumer, virtual machine
security requirement of resource of critical application and and cloud broker in details. [8]
data .Decision is made on the basis of rule. Performance of
resource manager is also evaluated by using cloudsim on basis The agent-based adaptive resource allocation is discussed in
of resource utilization and cost in sky environment. Set 2011 by the Gihun Jung, Kwang Mong Sim. In this paper the
priorities request and allocate resource accordingly. Sky provider needs to allocate each consumer request to an
computing provides computing concurrent access to multiple appropriate data center among the distributed data centers
clouds according user requirement. Define the Cloud services because these consumers can satisfy with the service in terms
like Software as a service (SaaS), Platform as a Service of fast allocation time and execution response time. Service
(PaaS) and Infrastructure as a service. [2] provider offers their resources under the infrastructure as a
service model. For IaaS the service provider delivers its
In 2010 Zhang Yu Hua, Zhang Jian ,Zhang Wei Hua present resources at the request of consumers in the form of VMs. To
argumentation about the intelligent cloud computing system find an appropriate data center for the consumer request,
and Data warehouse that record the inside and outside data of propose an adaptive resource allocation model considers both
Cloud Computing System for data analysis and data mining. the geographical distance between the location of consumer
Management problem of CCS are: balance between capacity and datacenters and the workload of data center. With
and demand, capacity development planning, performance experiment the adaptive resource allocation model shows
optimization, system safety management. Architecture of the higher performance. An agent based test bed designed and
Intelligence cloud computing system is defined with Data implemented to demonstrate the proposed adaptive resource
source, data warehouse and Cloud computing management allocation model. The test bed implemented using JAVA with
information system. [3] JADE (Java Agent Development framework). [9]
www.ijcat.com 779
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014
4. CONCLUSION
The user’s usages have large number of progress in an
environment. So there have large number of problems are
occurred in the cloud. The resource provisioning can be
overcome by hybrid method. This proposed method is used to
allocate the resources with working conditions. It shows the
energy is very efficiency and the overcome the workload with
the good performance.
www.ijcat.com 780
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 774 - 777, 2014
6. REFERENCES
[12]O. Duchenne, I. Laptev, J. Sivic, F. Bach, and J. Ponce.
[1]http://www.youtube.com/yt/press/statistics.html Au- tomatic annotation of human actions in video. In Proc. of
[2]http://nlp.stanford.edu/software/corenlp.shtml ICCV, 2009.
[3]Collins English Dictionary, entry for "lemmatise" [13]Laptev, M. Marszalek, C. Schmid, and B.
[4]L. Ratinov and D. Roth, Design Challenges and Rozenfeld.Learning realistic human actions from movies. In
Misconceptions in Named Entity Recognition. CoNLL (2009) Proc. of CVPR, 2008
[14]M. Everingham, J. Sivic, and A. Zisserman. Hello! my
[5]G. A. Miller.Wordnet: A lexical database for english.
name is... buffy automatic naming of characters in tv video. In
(11):39-41. Proc. of BMVC, 2006.
[6]Chengde Zhang, Xiao Wu, Mei-Ling Shyu and QiangPeng,
" Adaptive Association Rule Mining for Web Video Event [15]F. Smeaton, P. Over, and W. Kraaij.Evaluation campaigns
Classification", 2013 IEEE 14th International Conference on and trecvid. In Proc. of ACM Workshop on Multimedia In-
Information Reuse and Integration (IRI), page 618-625. formation Retrieval, 2006
[7] Y. Song, M. Zhao, J. Yagnik, and X. Wu.Taxonomic [16]J. Yang, R. Yan, and A. G. Hauptmann. Cross-domain
classification for web-based videos.In CVPR, 2010. video concept detection using adaptive svms. In Proc. of
[8] Z. Wang, M. Zhao, Y. Song, S. Kumar, and B. Li. ACM MM, 2007.
Youtube-cat: Learning to categorize wild web videos. In [17] M. E. Sargin, H. Aradhye, P. J. Moreno, and M. Zhao.
CVPR, 2010. Au- diovisual celebrity recognition in unconstrained web
[9] http://www.ranks.nl/resources/stopwords.html videos. In Proc. of ICASSP, 2009.
[10]http://cs.nyu.edu/grishman/jet/guide/PennPOS.html [18] J. Liu, J. Luo, and M. Shah.Recognizing realistic actions
[11]Roth and D. Zelenko, Part of Speech Tagging Using a from videos.In Proc. of CVPR, 2009.
Network of Linear Separators. Coling-Acl, The 17th [19] S. Zhang, C. Zhu, J. K. O. Sin, and P. K. T. Mok, "A
International Conference on Computational Linguistics (1998) novel ultrathin elevated channel low-temperature poly-Si
pp. 1136—1142 TFT," IEEE Electron Device Lett., vol. 20, pp. 569-571, Nov.
1999
www.ijcat.com 781
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 782 - 784, 2014, ISSN:- 2319–8656
Abstract: Mobile nodes in military environments such as a battlefield or a hostile region are likely to suffer from intermittent network
connectivity and frequent partitions. Disruption-tolerant network (DTN) technologies are becoming successful solutions that allow
wireless devices carried by soldiers to communicate with each other and access the confidential information or command reliably by
exploiting external storage nodes. However, the problem of applying CP-ABE in decentralized DTNs introduces several security and
privacy challenges with regard to the attribute revocation, key escrow, and coordination of attributes issued from different authorities.
In this paper, we propose a secure data retrieval scheme using CP-ABE for decentralized DTNs where multiple key authorities manage
their attributes independently. We demonstrate how to apply the proposed mechanism to securely and efficiently manage the
confidential data distributed in the disruption-tolerant military network. Since some users may change their associated attributes at
some point (for example, moving their region), or some private keys might be compromised, key revocation (or update) for each
attribute is necessary in order to make systems secure. This implies that revocation of any attribute or any single user in an attribute
group would affect the other users in the group. It may result in bottleneck during rekeying procedure, or security degradation due to
the windows of vulnerability if the previous attribute key is not updated immediately.
www.ijcat.com 782
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 782 - 784, 2014, ISSN:- 2319–8656
disadvantage of this fully distributed approach is the access from the storage node or key authorities should be also
performance degradation. Since there is no centralized prevented.
authority with master secret information, all attribute
authorities should communicate with each other in the system 3.2.2 Collusion-resistance:
to generate a user’s secret key. If multiple users collude, they may be able to decrypt a
ciphertext by combining their attributes even if each of the
users cannot decrypt the ciphertext alone.
3. SYSTEM DESIGN
3.2.3 Backward and forward Secrecy
3.1 Existing System In the context of ABE, backward secrecy means that any user
When multiple authorities manage and issue attribute keys to who comes to hold an attribute (that satisfies the access
users independently with their own master secrets, it is very policy) should be prevented from accessing the plaintext of
hard to define fine-grained access policies over attributes the previous data exchanged before he holds the attribute. On
issued from different authorities. the other hand, forward secrecy means that any user who
drops an attribute should be prevented from accessing the
plaintext of the subsequent data exchanged after he drops the
The problem of applying the ABE to DTNs introduces several attribute, unless the other valid attributes that he is holding
security and privacy challenges. Since some users may satisfy the access policy.
change their associated attributes at some point (for example,
moving their region), or some private keys might be
compromised, key revocation (or update) for each attribute is
necessary in order to make systems secure. However, this Please use a 9-point Times Roman font, or other Roman font
issue is even more difficult, especially in ABE systems, since with serifs, as close as possible in appearance to Times
each attribute is conceivably shared by multiple users Roman in which these guidelines have been set. The goal is to
(henceforth, we refer to such a collection of users as an have a 9-point text, as you see here. Please use sans-serif or
attribute group) non-proportional fonts only for special purposes, such as
.
Another challenge is the key escrow problem. In CP-ABE, the
key authority generates private keys of users by applying the 4. SYSTEM IMPLEMENTATION
authority’s master secret keys to users’ associated set of
attributes. The last challenge is the coordination of attributes 4.1 Key Authorities
issued from different authorities. When multiple authorities They are key generation centers that generate public/secret
manage and issue attributes keys to users independently with parameters for CP-ABE. The key authorities consist of a
their own master secrets, it is very hard to define fine-grained central authority and multiple local authorities. We assume
access policies over attributes issued from different that there are secure and reliable communication channels
authorities. between a central authority and each local authority during the
initial key setup and generation phase. Each local authority
manages different attributes and issues corresponding attribute
3.2 Proposed System keys to users.
First, immediate attribute revocation enhances
They grant differential access rights to individual users based
backward/forward secrecy of confidential data by reducing the
on the users’ attributes. The key authorities are assumed to be
windows of vulnerability.
honest-but-curious. That is, they will honestly execute the
assigned tasks in the system; however they would like to learn
Second, encryptors can define a fine-grained access policy
information of encrypted contents as much as possible.
using any monotone access structure under attributes issued
from any chosen set of authorities.
4.2 Storage node:
Third, the key escrow problem is resolved by an escrow-free This is an entity that stores data from senders and provide
key issuing protocol that exploits the characteristic of the corresponding access to users. It may be mobile or static.
decentralized DTN architecture. The key issuing protocol Similar to the previous schemes, we also assume the storage
generates and issues user secret keys by performing a secure node to be semi-trusted that is honest-but-curious.
two-party computation (2PC) protocol among the key
authorities with their own master secrets. The 2PC protocol
deters the key authorities from obtaining any master secret
4.3 Sender:
information of each other such that none of them could This is an entity who owns confidential messages or data
generate the whole set of user keys alone. (e.g., a commander) and wishes to store them into the external
data storage node for ease of sharing or for reliable delivery to
Thus, users are not required to fully trust the authorities in users in the extreme networking environments. A sender is
order to protect their data to be shared. The data responsible for defining (attribute based) access policy and
confidentiality and privacy can be cryptographically enforced enforcing it on its own data by encrypting the data under the
against any curious key authorities or data storage nodes in policy before storing it to the storage node.
the proposed scheme. 4.4 User
This is a mobile node who wants to access the data stored at
3.2.1 Data confidentiality: the storage node (e.g., a soldier). If a user possesses a set of
attributes satisfying the access policy of the encrypted data
Unauthorized users who do not have enough credentials defined by the sender, and is not revoked in any of the
satisfying the access policy should be deterred from accessing attributes, then he will be able to decrypt the ciphertext and
the plain data in the storage node. In addition, unauthorized obtain the data.
www.ijcat.com 783
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 782 - 784, 2014, ISSN:- 2319–8656
www.ijcat.com 784
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 785 - 790, 2014, ISSN:- 2319–8656
Abstract: Recently, much research in software engineering focused on reverse engineering of software systems which has become one
of the major engineering trends for software evolution. The objective of this survey paper is to provide a literature review on the
existing reverse engineering methodologies and approaches for documenting the architecture of software systems. The survey process
was based on selecting the most common approaches that form the current state of the art in documenting software architectures. We
discuss the limitations of these approaches and highlight the main directions for future research and describe specific open issues for
research.
Keywords: Reveres Engineering; Software Architecture; Documenting Software Architectures; Architectural Design Decisions.
www.ijcat.com 785
International Journal of Computer Applications Technology and Research
Volume 3* Issue 12, December 2014
www.ijcat.com 786
International Journal of Computer Applications Technology and Research
Volume 3* Issue 12, December 2014
www.ijcat.com 787
International Journal of Computer Applications Technology and Research
Volume 3* Issue 12, December 2014
Author Problem
# Proposed Solution(s) Results and Findings Limitation(s)
(year) Statement
2 Hugo et al Understanding the - Generic and extensible - MoDisco provided MoDisco should extend to
(2014) contents of the MDRE framework high adaptability support additional technologies
legacy systems called "MoDisco". because it is based on and include more advanced
using model the nature of legacy components to improve the
driven reverse - This framework is system technologies system comprehension.
engineering applicable to different and the scenario(s)
(MDRE) types of legacy based on reverse
systems. engineering.
3 Che et al Collecting - Triple View Model - TVM framework TVM framework should extend
(2011) architectural (TVM) an architecture includes three to manage the evaluation and
design decisions framework for different views for documentation of ADDs by
(ADDs) documenting ADDs. describing the specifying its views through the
notation of ADDs. stakeholders' scenarios.
4 Che et al Managing the - Scenario based - Manage the There is a need to support
(2012) documentation method (SceMethod) documentation and multiple ways on managing and
and evolution of for documenting and the evaluation needs documenting the ADDs during
the architectural evaluating ADDs. for ADDs through the architecture process.
design decisions stakeholders'
- This solution is based scenario(s).
on TVM. The main
purpose is to apply
TVM for specifying
its views through end-
user scenario(s).
www.ijcat.com 788
International Journal of Computer Applications Technology and Research
Volume 3* Issue 12, December 2014
Author Problem
# Proposed Solution(s) Results and Findings Limitation(s)
(year) Statement
5 Che Documenting - Developed UML - Apply the evaluation This solution is focused on the
(2013) and evolving the Metamodel for the TVM related to the developers view point and
architectural framework. The main specified attributes their work is currently in
design decisions purpose was to make and establish progress to support the ADD
each view of TVM traceable evaluation documentation and evaluation
specified by classes and a of ADDs, in geographically separated
set of attributes for software development (GSD).
- Allow explicit
describing ADDs
evaluation knowledge
information.
of ADDs.
- Support multiple
ways for
documenting ADDs
during the
architecture process.
6 Shahin et A survey of - The purpose of this - All of selected ADD There is a need to focus on
al (2009) architectural survey was to investigate models include the stakeholder to communicate
design decision ADD models to decide if major elements. with the stored knowledge
models and tools there are any similar of ADDs. This could be
- Most of ADD models
concepts or differences achieved by applying the
are based on using
on capturing ADD. scenario based
different minor
documentation and
- The survey classified elements for
evaluation methods through
ADD concept into two capturing and
stakeholders' scenario(s) to
categories: Major documenting the
manage the documentation
elements which refer to ADD.
and the evaluation needs for
the consensus on - All of selected ADD ADDs.
capturing and models deal with the
documenting ADD based architecture design as
on the constraint, the decision making
rationale and alternative process.
of decision. While the
Minor elements refers to - Not all models were
the elements that used supported by tools.
without consensus on Hence, some of these
capturing and ADD based on text
documenting ADD. template for
capturing and
- Moreover, to clarify the documenting ADDs.
desired features that are - However, most of
missed according to the existing ADD tools
architecture needs do not support the
ability of
stakeholders to
communicate with
the stored knowledge
of ADD.
www.ijcat.com 789
International Journal of Computer Applications Technology and Research
Volume 3* Issue 12, December 2014
www.ijcat.com 790
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
Abstract: Data center have the largest consumption amounts of energy in sharing the power. The public cloud workloads of different
priorities and performance requirements of various applications [4]. Cloud data center have capable of sensing an opportunity to present
different programs. In my proposed construction and the name of the security level of imperturbable privacy leakage rarely distributed
cloud system to deal with the persistent characteristics there is a substantial increases and information that can be used to augment the
profit, retrenchment overhead or both. Data Mining Analysis of data from different perspectives and summarizing it into useful
information is a process. Three empirical algorithms have been proposed assignments estimate the ratios are dissected theoretically and
compared using real Internet latency data recital of testing methods.
www.ijcat.com 791
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
www.ijcat.com 792
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
the Cloud infrastructure and consumers. The interactions of for data transmission client request processing physical
the following components to support energy efficient resource machine data population server mining user identification by
management. The components are Arbitrator, Overhaul- server using the request name.
Prediction, Client Focused, Budget Processing, Power 4.5 Efficient Server Provisioning
Consuming, Overhaul-Router, Storage Leadership and Gauge. Dynamic allocation has been user request to the physical
server in done by cloud environment by using the concept of
protocol named as gossip [5]. This protocol is sufficient
protocol for dynamic resource allocation and it gives response
to the client at exact query matching server provisioning
approach for minimized the cost, reduce the time and quick
response. In order to assemble the cloud environmental setup
and physical server storage device is very expensive but they
are applying the mining setup. Show that it is must like
expensive.
5. EXPERIMENTAL RESULT
The implementation of my concept is to create the n number
of virtual machines and physical machines. In this machines
have n number of information’s are stored. This physical
machine contains java connection based classes and service
based concepts. In cloud environment system distributed set
up of mined servers. The cloud server’s shows all the
information of the physical server in the data center.
Figure: 3.1 Green Cloud Computing Architecture
In data center having mining setup for retrieve the data from
the data storage. In data storage wants to store the number of
Gossip huge clouds of green computing resource allocation
files with the help of query processing from server. In this
based they aim to reduce server power consumption by method using SVM for classify the data for the user query
integration with the specific objectives are sensible the searched from the server.
resource allocation protocol and propose a common rumors.
The data center is the large storage network. The network
Under load the ethical and fair allocation of CPU resources to wants to secure the stored information for the storage devices
clients. The simulation result and the key performance of by using the cryptography technique. In this concept am using
metrics for resource allocation process and suggest that it is MD5 method for creating number of keys to secure the data
the body has do not change with increasing size. from the storage devices. The valid user’s are only views the
information about the user queries. The key based concept is
4. SYSTEM IMPLEMENTATION achieving the more secure for storage devices.
4.1 General Finally, the server client communication is the very large
In general, user interface design for consumer requests and process. In this having number of resources and it also
getting response from the server. The consumer is valid intermediate process for file transfers to the user’s and data
provider means they are getting the further processes. It is storage. It is the nonstop processes so here using gossip
used to interact with client and server for request and protocol for green computing process. It is an automatically
responding from the cloud storage. allocate the resources for file sharing in the data center.
www.ijcat.com 793
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 791 - 794, 2014, ISSN:- 2319–8656
process for user’s from the storage devices. In this device [4] Q. Zhang, M.F. Zhani, R. Boutaba, and J.L. Hellerstein,
having continue processes so the machines want to overcome “HARMONY: Dynamic Heterogeneity-Aware Resource
the delay processing by using the method of gossip protocol Provisioning in the Cloud,”Proc. IEEE Int’l Conf.
for green computing. Distributed Computing Systems (ICDCS), 2013.
[5] Q. Zhang, M.F. Zhani, Q. Zhu, S. Zhang, R. Boutaba,
and J.L. Hellerstein, “Dynamic Energy-Aware Capacity
7. REFERENCES Provisioning for Cloud Computing Environments,” Proc.
[1] Qi Zhang, Mohamed Faten Zhani, Raouf Boutaba, ACM Int’l Conf. Autonomic Computing (ICAC), 2012.
Joseph L. Hellerstein “Dynamic Heterogeneity Aware
Resource Provisioning in the Cloud” Proc.IEEE Trans. [6] Lu Zhang, Xueyan Tang, “Optimizing Client Assignment
Cloud computing,2014. for Enhancing Interactivity in Distributed Interactive
Applications,” Proc. IEEE/ACM Transaction on
[2] Jun Zhou, Xiaodong Lin, Xiaolei Dong, Zhenfu Cao Networking, 2012.
“PSMPA: Patient Self-controllable and Multi-level
Privacy-preserving Cooperative Authentication in [7] P. Morillo, J. Orduna, M. Fernandez, and J. Duato,
Distributed m-Healthcare Cloud Computing System” “Improving the performance of distributed virtual
Proc. IEEE Tans. Parallel and Distributed System, 2014. environment systems,” IEEE Trans. Parallel Distrib.
Syst., vol. 16, no. 7, pp. 637–649, Jul. 2005.
[3] Lu Zhang and Xueyan Tang “The Client Assignment
Problem for Continuous Distributed Interactive [8] J. Sun and Y. Fang, Cross-domain Data Sharing in
Applications: Analysis, Algorithms, and Evaluation” Distributed Electronic Health Record System, IEEE
Proc. IEEE Trans. Parallel and Distributed System, 2014. Transactions on Parallel and Distributed Systems, vol.
21, No. 6, 2010.
Authors
Dr.T.Senthil Prakash received the Ph.D. degree from the PRIST University,
Thanjavur, India in 2013 and M.E(CSE) degree from Vinayaka Mission’s University,
Salem, India in 2007 and M.Phil.,MCA.,B.Sc(CS) degrees from Bharathiyar
University, Coimbatore India, in 2000,2003 and 2006 respectively, all in Computer
Science and Engineering. He is a Member in ISTE New Delhi, India, IAENG, Hong
Kong..IACSIT, Singapore SDIWC, USA. He has the experience in Teaching of
10+Years and in Industry 2 Years. Now He is currently working as a Professor and
Head of the Department of Computer Science and Engineering in Shree Venkateshwara
Hi-Tech Engineering College, Gobi, Tamil Nadu, and India. His research interests
include Data Mining, Data Bases, Artificial Intelligence, Software Engineering etc.,He
has published several papers in 17 International Journals, 43 International and National
Conferences.
www.ijcat.com 794
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 795 - 798, 2014, ISSN:- 2319–8656
Abstract: Generally, the botnet is one of the most dangerous threats in the network. It has number attackers in the network. The
attacker consists of DDOS attack, remote attack, etc., Bots perform perform repetitive tasks automatically or on a schedule over the
internet, tasks that would be too mundane or time-consuming for an actual person. But the botnets have stealthy behavior as they are
very difficult to identify. These botnets have to be identified and the internet have to be protected. Also the the activity of botnets must
be prevented to provide the users, a reliable service. The past of botnet detection has a transaction process which is not secure. A
efficient stastical data classifier is required to train the botent preventions system. To provide the above features clustering based
analysis is done. our approach can detect and profile various P2P applications rather than identifying a specific P2P application.
Anomaly based detection technique is used to obtain this goal.
Keywords: Botnet, anomaly base detection, hash function, DDOS
www.ijcat.com 795
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 795 - 798, 2014, ISSN:- 2319–8656
exploiting, etc.[4] Unfortunately, the malicious activities may sets of flows Stcp(h) and Sudp(h), we characterize each flow
be stealthy and non-observable A efficient stastical data using a vector of statistical features v(h) = [Pkts , Pktr , Bytes
classifier is required to train the botent preventions system. , Byter ], in which Pkts and Pktr represent the number of
Acquiring such information is a challenging task, thereby packets sent and received, and Bytes and Byter represent the
drastically limiting the practical use of these methods. Some number of bytes sent and received, respectively.
of the older approach involves content signature, encryptions,
profiling, fixed source port. our approach does not need any The distance between two flows is subsequently defined as
content signature. our analysis approach can estimate the the euclidean distance of their two corresponding vectors. We
active time of a P2P application, which is critical for botnet then apply a clustering algorithm to partition the set of flows
detection into a number of clusters. Each of the obtained clusters of
flows, Cj (h), represents a group of flows with similar size.
3. SYSTEM DESIGN
A Botmaster has to be designed with P2P protocol. Flows corresponding to ping/pong and peer-discovery share
Therefore P2P bots exhibit some network traffic patterns that similar sizes, and hence they are grouped into two clusters
are common to other P2P client applications either legitimate (FC1 and FC2), respectively. Since the number of destination
or malicious. Hence our system is divided into two phases. In BGP prefixes involved in each cluster is larger, we take FC1
the first phase, we aim at detecting all hosts within the and FC2 as its fingerprint clusters. A fingerprint cluster
monitored network that engage in P2P communications. We summary, (Pkts , Pktr , Bytes , Byter , proto), represents the
analyze raw traffic collected at the edge of the monitored protocol and the average number of sent/received
network and apply a pre-filtering step to discard network packets/bytes for all the flows in this fingerprint cluster. We
flows that are unlikely to be generated by P2P[1]. We then implemented the flow analysis component and identified
analyze the remaining traffic and extract a number of fingerprint cluster for the sample P2P traces including two
statistical features to identify flows generated by P2P clients. traces.
In the second phase, our system analyzes the traffic generated
by the P2P clients and classifies them into either legitimate 3.2 Detecting P2P bots
P2P clients or P2P bots. Specifically, we investigate the active To detect the bots coarse grained detection method is used.
time of a P2P client and identify it as a candidate P2P bot if it Since bots are malicious programs used to perform profitable
is persistently active on the underlying host. We further malicious activities, they represent valuable assets for the
analyze the overlap of peers contacted by two candidate P2P botmaster, who will intuitively try to maximize utilization of
bots to finalize detection. After analyzing with the use of bots. This is particularly true for P2P bots[5] because in order
anomaly based detection algorithm the network has to be to have a functional overlay network (the botnet), a sufficient
revoked from malwares. number of peers needs to be always online. In other words,
the active time of a bot should be comparable with the active
time of the underlying compromised system.
Network traffic
Traffic filter The distance between each pair of hosts is computed. We
apply hierarchical clustering, and group together hosts
according to the distance defined above. In practice the
hierarchical clustering algorithm will produce a dendrogram
Detecting Detecting P2P (a tree-like data structure). The dendrogram expresses the
Bots P2P Bots clients “relationship” between hosts. The closer two hosts are, the
lower they are connected at in the dendrogram. Two P2P bots
in the same botnet should have small distance and thus are
connected at lower level. In contrast, legitimate P2P
Revoking from applications tend to have large distances and consequently are
malware connected at the upper level. We then classify hosts in dense
clusters as P2P bots, and discard all other clusters and the
related hosts, which we classify as legitimate P2P clients.
Fig 1: System architecture
4. SYSTEM IMPLEMENTATION
3.1 Detecting P2P client Out of four components in our system, “Traffic Filter” and
Traffic filter is used to sort out th traffic which is unlikely to “Coarse-Grained Detection of P2P Bots” have linear
P2P networks. In this first phase, fine grained detection of complexity since they need to scan flows only once to identify
P2P botenets is implemented. This component is responsible flows with destination addresses resolved from DNS queries
for detecting P2P clients by analyzing the remaining network or calculate the active time. Other two components, “Fine-
flows after the Traffic Filter component. For each host h Grained Detection of P2P Clients” and “Fine-Grained P2P
within the monitored network we identify two flow sets, Detection of P2P Bots”, require pairwise comparison for
denoted as Stcp(h) and Sudp(h), which contain the flows distance calculation
related to successful outgoing TCP and UDP[6] connection,
respectively. We use a two-step clustering approach to reduce the time
complexity of “Fine-Grained P2P Client Detection”. For the
To identify flows corresponding to P2P control messages,we first-step clustering, we use an efficient clustering algorithm
first apply a flow clustering process intended to group to aggregate network flows into K sub-clusters, and each
together similar flows for each candidate P2P node h. Given subcluster contains flows that are very similar to each other.
For the second-step clustering, we investigate the global
www.ijcat.com 796
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 795 - 798, 2014, ISSN:- 2319–8656
distribution of sub-clusters and further group similar sub- scenario always visible botnet minor. If any dispute in the
clusters into clusters. identification scenario overall network may be crashed.
The distance of two flows is defined as the Euclidean distance 4.1.4 Attacking model of Malware
of their corresponding vectors, where each vector [Pkts , Pktr , Botnet minor contain all the details about the peer network.
Bytes , Byter ] represents the number of packets/ bytes that The botnet minor handles all the request processed by the
are sent/received in a flow. decentralized network. The botnet major attack decentralized
scenario spread the warm data to the peer network. The node
For the second-step clustering, we use hierarchical clustering connected with the attacked node that specific node also get
with DaviesBouldin validation[8] to group sub-clusters into the warm data.
clusters. Each sub-cluster is represented using a vector ([Pkts ,
Pktr , Bytes , Byter ]), which is essentially the average for all
flow vectors in this sub-cluster. 4.1.5 Revoking the network from Malware
Data matching have the law data and the original data. The
Hierarchical clustering is used to build a dendrogram. Finally, proposed technical approach can identify the warm data it is
DaviesBouldin validation is employed to assess the global spreaded by the botnet. Revoke the original data instead of
distribution of inter- and intra-cluster distances of clusters warm data it can identify the problem and revoke the botnet
based on various clustering decisions and yield the best cut for minor from the attacking model.
the dendrogram. The two-step clustering algorithm has the
time complexity of O(nK I + K2).
5. EXPERIMENTAL RESULTS
4.1 Modules We prepared a data set (D) for evaluation. Specifically, we
The goal of guarding the large scale scrabble in social randomly selected half [8] of the P2P bots from NETbots
network is implemented by the following modules, .Then for each of the 5 P2P applications we ran, we randomly
selected one out of its two traces from NETP2P and overlaid
its traffic to the traffic of a randomly selected host We applied
User interface design our detection system on data set D. The traffic filter
drastically reduced the workload for the whole system. As
indicated in Figure 4, it reduced the number of hosts subject to
analysis by 67% (from 953 to 316) but retained all P2P
clients.
Implementing peer netwrok
Among 26 P2P clients identified in the previous step, 25 out
of them exhibit persistent P2P behaviors. We further evaluate
the similarity of fingerprint clusters and peer IPs for each pair
of persistent P2P clients and
Botnet minor approach
derive a dendrogram.
6. CONCLUSION
Revoking from malware To summarize, although our system greatly enhances and
complements the capabilities of existing P2P botnet detection
systems, it is not perfect. We should definitely strive to
4.1.1 User Interface Design develop more robust defense techniques, where the
The user interaction is effective operation and control of the aforementioned discussion outlines the potential
machine on the user„s. The user interface module has login improvements of our system.
and registration phases. The registration phase gets details
from user and stores it in database. It also checks the details In this paper, we presented a novel botnet detection system
are valid or not. that is able to identify stealthy P2P botnets, whose malicious
activities may not be observable. To accomplish this task, we
derive statistical fingerprints of the P2P communications to
4.1.2 Implementing peer network first detect P2P clients and further distinguish between those
The peer network contain decentralized networks. All the that are part of legitimate P2P networks (e.g., filesharing
nodes contains separate IP address and separate port number. networks) and P2P bots. We also identify the performance
The peer one node have stored separate list of files which in bottleneck of our system and optimize its scalability. The
the global respository. evaluation results demonstrated that the proposed system
accomplishes high accuracy on detecting stealthy P2P bots
4.1.3 Botnet minor approach and great scalability.
The global respository contains the decentralized network
details. The botnet minor store and reteive the information
about port and IP details from the database. Identification
www.ijcat.com 797
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 795 - 798, 2014, ISSN:- 2319–8656
http://mtc.sri.com/Conficker/addendumC/index.html
Ms. Geerthidevi K G, PG Scholar Currently persuing her M.E CSE degree in Shree Venkateshwara Hi-
Tech Engg College, Gobi, Tamilnadu , India. Her research interests include Networking, Network Security
etc.,
Dr.T.Senthil Prakash received the Ph.D. degree from the PRIST University, Thanjavur, India in 2013
and M.E(CSE) degree from Vinayaka Mission‟s University, Salem, India in 2007 and
M.Phil.,MCA.,B.Sc(CS) degrees from Bharathiyar University, Coimbatore India, in 2000,2003 and 2006
respectively, all in Computer Science and Engineering. He is a Member in ISTE New Delhi, India,
IAENG, Hong Kong..IACSIT, Singapore SDIWC, USA. He has the experience in Teaching of 10+Years
and in Industry 2 Years. Now He is currently working as a Professor and Head of the Department of Computer Science
and Engineering in Shree Venkateshwara Hi-Tech Engineering College, Gobi, Tamil Nadu, and India. His research
interests include Data Mining, Data Bases, Artificial Intelligence, Software Engineering etc.,He has published several
papers in 17 International Journals, 43 International and National Conferences.
www.ijcat.com 798
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
Abstract: Cloud Computing is going to be very popular technology in IT enterprises. For any enterprise the data stored is very huge
and invaluable. Since all tasks are performed through network it has become vital to have the secured use of legitimate data. In cloud
computing the most important matter of concern are data security and privacy along with flexibility, scalability and fine grained access
control of data being the other requirements to be maintained by cloud systems Access control is one of the prominent research topics
and hence various schemes have been proposed and implemented. But most of them do not provide flexibility, scalability and fine
grained access control of the data on the cloud. In order to address the issues of flexibility, scalability and fine grained access control
of remotely stored data on cloud we have proposed the hierarchical attribute set-based encryption (HASBE) which is the extension of
attribute- set-based encryption(ASBE) with a hierarchical structure of users. The proposed scheme achieves scalability by handling the
authority to appropriate entity in the hierarchical structure, inherits flexibility by allowing easy transfer and access to the data in case
of location switch. It provides fine grained access control of data by showing only the requested and authorized details to the user thus
improving the performance of the system. In addition, it provides efficient user revocation within expiration time, request to view
extra-attributes and privacy in the intra-level hierarchy is achieved. Thus the scheme is implemented to show that is efficient in access
control of data as well as security of data stored on cloud with comprehensive experiments.
www.ijcat.com 799
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
www.ijcat.com 800
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
policies into user‟s keys; while in our system attributes are hierarchical identity-based encryption (HIBE) system and a
used to describe a user‟s credentials, and a party encrypting ciphertext-policy attribute-based encryption (CP-ABE) system,
data deter-mines a policy for who can decrypt. Thus, our meth- so as to provide not only fine-grained access control, but also
ods are conceptually closer to traditional access control full delegation and high performance. Then, we propose a
methods such as Role-Based Access Control (RBAC). In scalable revocation scheme by applying proxy re-encryption
addition, we provide an implementation of our sys- tem and (PRE) and lazy re-encryption (LRE) to the HABE scheme, so
give performance measurements. as to efficiently revoke access rights from users.
2.5 Ayad et al. “Enabling Data Dynamic and 2.7 Qin et al. “Hierarchical Attribute-Based
Indirect Mutual Trust for Cloud Computing Encryption for Fine-Grained Access Control
Storage System”, 2012. [6] in Cloud Storage Services”. [9]
In this paper, we propose a cloud-based storage scheme that Cloud computing, as an emerging computing paradigm,
allows the data owner to benefit from the facilities offered by enables users to remotely store their data into a cloud so as to
the CSP and enables indirect mutual trust between them. The enjoy scalable services on-demand. Especially for small and
proposed scheme has four important features: (i) it allows the medium-sized enterprises with limited budgets, they can
owner to outsource sensitive data to a CSP, and perform full achieve cost savings and productivity enhancements by using
block-level dynamic operations on the outsourced data, i.e., cloud-based services to manage projects, to make
block modification, insertion, deletion, and append, (ii) it collaborations, and the like. However, allowing cloud service
ensures that authorized users (i.e., those who have the right to providers (CSPs), which are not in the same trusted domains as
access the owner‟s file) receive the latest version of the enterprise users, to take care of confidential data, may raise
outsourced data, (iii) it enables indirect mutual trust between potential security and privacy issues. To keep the sensitive user
the owner and the CSP, and (iv) it allows the owner to grant or data confidential against untrusted CSPs, a natural way is to
revoke access to the outsourced data. We discuss the security apply cryptographic approaches, by disclosing decryption keys
issues of the proposed scheme. Besides, we justify its only to authorized users. However, when enterprise users
performance through theoretical analysis and experimental outsource confidential data for sharing on cloud servers, the
evaluation of storage, communication, and computation adopted encryption system should not only support fine-
overheads. grained access control, but also provide high performance, full
delegation, and scalability, so as to best serve the needs of
2.6 Guojun et al. “Hierarchical attribute- accessing data anytime and anywhere, delegating within
based encryption and scalable user enterprises, and achieving a dynamic set of users. In this paper,
we propose a scheme to help enterprises to efficiently share
revocation for sharing data in cloud
confidential data on cloud servers. We achieve this goal by first
servers”, 2011. [8] combining the hierarchical identity-based encryption (HIBE)
system and the cipher text-policy attribute-based encryption
With rapid development of cloud computing, more and more
(CP-ABE) system, and then making a performance-
enterprises will outsource their sensitive data for sharing in a
expressivity tradeoff, finally applying proxy re-encryption and
cloud. To keep the shared data confidential against untrusted
lazy re-encryption to our scheme.
cloud service providers (CSPs), a natural way is to store only
the encrypted data in a cloud. The key problems of this 2.8. Patrick et al. “Methods and Limitations
approach include establishing access control for the encrypted
of Security Policy Reconciliation”. [10]
data, and revoking the access rights from users when they are
no longer authorized to access the encrypted data. This paper
A security policy is a means by which participant session
aims to solve both problems. First, we propose a hierarchical
requirements are specified. However, existing frameworks
attribute-based encryption scheme (HABE) by combining a
provide limited facilities for the automated reconciliation of
www.ijcat.com 801
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
participant policies. This paper considers the limits and 3. This system is designed to provide flexibility of the
methods of reconciliation in a general-purpose policy model. data where in case of transfer of employee, his data
We identify an algorithm for efficient two-policy could be transferred to respective location with ease.
4. It also provides scalability in case when an employee
reconciliation, and show that, in the worst-case, reconciliation
is absent his work could be handled by the senior
of three or more policies is intractable. Further, we suggest
employee securely.
Efficient heuristics for the detection and resolution of
intractable reconciliation. Based upon the policy model, we
describe the design and implementation of the Ismene policy
language. The expressiveness of Ismene, and indirectly of our
model, is demonstrated through the representation and
exposition of policies supported by existing policy languages.
We conclude with brief notes on the integration and
enforcement of Ismene policy within the Antigone.
3. PROPOSED SYSTEM
In our propose system instead of showing complete data from
cloud we are fetching only those data which is essential for that
user. We are not fetching all data so it takes less time for
Figure 1. General Architecture of the System
fetching data so system response time is very less due to which
system performance increases. We are performing encryption
before storing data so even if data get hack by hacker data
3.1 Methodology
1. Registration and login by user:
cannot be easily understand by hacker. We are performing
In this user fill his/her own complete data. Request is sent to
hierarchical structure so even if lower authority is absent for
the CEO for confirmation. CEO confirms his/her request and
particular days at that time higher authority handle all work of
assigns attribute and time period for that user. Once Account
lower authority so work of company will not be stopped. The
get confirm password and key is sent to that user by email so
HASBE scheme for realizing scalable, flexible and fine-
grained access control in cloud computing. The HASBE he/she can access his/her account.
scheme seamlessly incorporates a hierarchical structure of
system users by applying a delegation algorithm to ASBE. 2. Approve User and Assign attributes:
HASBE not only supports compound attributes due to flexible Out of the selected attributes according the roles defined in
attribute set combinations, but also achieves efficient user hierarchy of the system the attribute visibility access is
revocation because of multiple value assignments of attributes. decided. Each attribute is encrypted.
We formally proved the security of HASBE based on the
security of CP-ABE. Finally, we completed the detailed 3. Key Generation and Verification
analysis of proposed scheme, and conducted comprehensive
Key is generated based on the attributes filled by the user in
performance analysis and evaluation, which showed its
efficiency and advantages over existing schemes. registration form. In attribute key verification, when a key is
used for login ,it is first checked with the key stored in the
database. If a match is found then user is allowed for further
3.1 Project Scope process else the user is rejected for further process.
4. Encryption and decryption of data
1. This system is designed to provide security to data
User fills his/her data during registration. Once it is click on
stored on cloud and improve performance of system
by showing only the required details requested by an submit button data is send to encryption algorithm that are
employee. RSA and AES. After performing encryption data is stored in
2. Security is provided by generating a secret key from encrypted format in database.
the various attributes stated in the form which is
5. Access Right:
filled by the employee at the time of registration. The user can view the selected attributes of the same level as
well as other levels according to the access authority using
attribute key.
www.ijcat.com 802
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
6. Fine Grained Access The cloud server provides the six main functions to
In our propose system instead of showing complete data, the
the user.
fetching of necessary data is allowed. Due to this system
provides a quick response time. 1. Fine-Grained Access
In our propose system instead of showing complete data,
7. Request for extra attribute: the fetching of necessary data is allowed. Due to this
system provides a quick response time.
The user can access attributes of same level as inter level
counterparts. He can request for extra attributes in case of
2. Scalability
emergency as well as ease of work.
We are performing hierarchical structure so even if
8. Flexibility lower authority is absent for particular days at that time
In this module flexibility can be done by suppose user is higher authority handle all work of lower authority so
transfer from one location to another location and for that new work of company will not be stopped.
location that user‟s data is not accessible then authority request
for accessing data of that user from old location. Once 3. Flexibility
authority got request that data should be access from new When an employee gets transferred, his data could be
location and it is not visible for old location accessible to the branch where he will be transferred
only not to the older branch. So data will be transferred
9. Scalability:
on request of CEO safely. Hence data can be transferred
We are performing hierarchical structure so even if lower
authority is absent for particular days at that time higher easily between branches.
authority handle all work of lower authority so work of
company will not be stopped. 4. Encryption
Encryption is a process in which data is hidden in a way
10. Efficient User Revocation: that is accessible to the authorized user only. In this
It can be done by two steps request to the admin and response system we are providing encryption (converting into
to the user from admin within expiration time.
unreadable) so that data is not accessible by any illegal
user like a hacker.
9. Privacy:
5. Decryption
Default it is public but a user can set intra-level privacy by Decryption is a process in which encrypted data i.e
restricting access to attributes.
unreadable format is converted into readable format.
3.2 Process Summary
Following Processes will be involved in Project: 6. Key Generation and Verification
Key is generated based on the attributes filled by the user
in registration form. In attribute key verification, when a
1. Encrypt data before Insert
key is used for login, it is first checked with the key
After user click on submit button data encrypted using RSA
stored in the database. If a match is found then user is
and AESs algorithm. Once data get encrypted it is stored into
database and when user wants to retrieve data it again
allowed for further process else the user is rejected for
decrypted and shown in original form. further process.
www.ijcat.com 803
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
messages. Messages encrypted with the public key can only be Where „AT‟ is main set of registered Attribute like at1, at2,
decrypted in a reasonable amount of time using the private key. at3…
User1 transmits her public key (n, e) to User2 and keeps the RAA= {raa1, raa2, raa3}
private key secret. User1 then wishes to send message M to
User2.He first turns M into an integer m, such that 0 ≤ m < n Where „RAA‟ is main set of Request for another Attribute
by using an agreed-upon reversible protocol known as a raa1, raa2, raa3
scheme. He then computes the cipher text c corresponding to
D] Identify the employee requested for another employee
Information
This can be done quickly using the method of exponentiation REI= {rei1, rei2, rei3}
by squaring. User1 then transmits c to User2.Note that at least
nine values of m will yield a cipher text c equal to m, but this is Where „REI‟ is main set of Request for another Attribute rei1,
very unlikely to occur in practice. rei2, rei3
User can recover m from c by using her private key exponent. AK= {ak1, ak2, ak3….}
Given m, user can recover the original message M by reversing F] Identify the processes as P.
the padding scheme.
P= {Set of processes}
P = {P1, P2, P3,P4……}
4.1.2 Advanced Encryption Standard Algorithm
P1 = {e1, e2, e3}
Where
The AES algorithm is also used for improving the searching
{e1= upload data on server}
and access mechanism.
We are using NP-Complete because it gives output within fix {e3= get new attribute after request}
interval of time.
{e4= get new employee information when employee get
transfer.}
Set Theory Analysis
G] Identify failure cases as FL
A] Identify the Employees
Failure occurs when –
E= {e1, e2, e3….} FL= {F1, F2, F3…}
a) F1= = {f| ‟f‟ if error in uploading due to interrupted Internet
Where „E‟ is main set of Employees like e1, e2, e3…
connection}
H] Identify success case SS:-
B] Identify the Attribute
Success is defined as-
AT= {at1, at2, at3….} SS= {S1, S2, S3, S4}
www.ijcat.com 804
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
INITIAL STATE:
S1=A B
U={R, UR}
R=registered user
Hierarchy
UR=unregistered user
H = {H1, H2, H3, H4}
Where,
H is cloud
H1 is CEO.
H2 is general manager.
www.ijcat.com 805
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
U= {H1, H2}
U‟=absent user
FLEXIBILITY:
Where,
C3 is the new branch where employee got transferred to. Identify the processes as P.
P= {Set of processes}
P = {P1, P2, P3, P4……}
H
Where
P1 = {S1, S2, S3}
P1 P2 P3 P4
H= {C1, C2, C3}
Where,
Where,
S2= (C1-C2) U C3
www.ijcat.com 806
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
U=present user
H
U‟=absent user
FLEXIBILITY:
Where,
S3
C1 is the old branch of the company where employee worked
before transfer. FINAL STATE:
C3 is the new branch where employee got transferred to. P= {Set of processes}
P = {P1, P2, P3, P4……}
Where
H P1 = {S1, S2, S3}
P1 P2
P2 P3 P4
H= {C1, C2, C3}
Where,
H transfer.}
P1
S1 S2 S3
S2= (C1-C2) U C3
Scalability
U= {H1, H2}
www.ijcat.com 807
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 799 - 808, 2014, ISSN:- 2319–8656
www.ijcat.com 808
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 809 - 811, 2014, ISSN:- 2319–8656
Abstract - Website security is a critical issue that needs to be considered in the web, in order to run your online business healthy and
smoothly. It is very difficult situation when security of website is compromised when a brute force or other kind of attacker attacks on
your web creation. It not only consume all your resources but create heavy log dumps on the server which causes your website stop
working.
Recent studies have suggested some backup and recovery modules that should be installed into your website which can take timely
backups of your website to 3rd party servers which are not under the scope of attacker. The Study also suggested different type of
recovery methods such as incremental backups, decremental backups, differential backups and remote backup.
Moreover these studies also suggested that Rsync is used to reduce the transferred data efficiently. The experimental results show
that the remote backup and recovery system can work fast and it can meet the requirements of website protection. The automatic backup
and recovery system for Web site not only plays an important role in the web defence system but also is the last line for disaster
recovery.
This paper suggests different kind of approaches that can be incorporated in the WordPress CMS to make it healthy, secure and
prepared web attacks. The paper suggests various possibilities of the attacks that can be made on CMS and some of the possible
solutions as well as preventive mechanisms.
Some of the proposed security measures –
1. Secret login screen
2. Blocking bad boats
3. Changing db. prefixes
4. Protecting configuration files
5. 2 factor security
6. Flight mode in Web Servers
7. Protecting htaccess file itself
8. Detecting vulnerabilities
9. Unauthorized access made to the system checker
However, this is to be done by balancing the trade-off between website security and backup recovery modules of a website, as measures
taken to secure web page should not affect the user‟s experience and recovery modules.
www.ijcat.com 809
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 809 - 811, 2014, ISSN:- 2319–8656
panel into sleep mode so that no one can attack your login Guiyang ; Xiaoyao Xie analyses the network system
page. faced by common threats and attack methods and means
Some of the proposed security measures are as for the typical, sum-up a website security system by the
follows– need to address the problem, solve these problems
1. Security for user accounts formulate the corresponding protection measures.
2. Security for login module
3. Security while registering user
4. Security related to database module 4. SECURITY MEASURES
5. htaccess and configuration file backup and restore Security for user account
6. Functionality to blacklist and whitelist Sometimes CMS might have user account with default
7. Firewall protection and prevention of brute force user name „admin‟ which is easier for attacker to predict
login attack and attack or query to your CMS. It is considered as bad
8. whois lookup and security checker security practice as it makes the task of attacker 50% easier
9. Security for comment spam because attacker already knows one of your credentials
10. Disabling access to source code and selection of required to login. Besides this a Password strength tool can
text on UI be used to allow you to create very strong passwords.
Backup can be taken using different approaches such Security for login module
as incremental backup, selective backup, complete backup It is to protect the CMS against brute force login
and user can also recover from the hacking attack by using attacks with the login lockdown feature so that users with
the restore mechanism which will restore system to certain IP range can be locked out of the system for a
previous working state. Backup can be complete database predetermined amount of time based on the configuration
backup. setting. It also force logout of all users after a configured
This paper basically deals with mechanisms period of time
mentioned above to secure website from bad boats and
hackers and make your server healthy by removing 1. Security while registering user
possible security threats. The Paper also pesents different Enable manual approve feature for registered
backup and restore mechanisms. accounts can minimize spam and bogus
registrations. Captcha can also help us to prove
valid user.
2. RELATED WORK
2. Security related to database module
There has been extensive efforts made to understand Table prefixes can be modified to other
web security by considering network traffic, encryption prefixes to make security level higher for the
techniques etc. But very few efforts have been taken to attacker. Attacker cannot predict the table prefix
understand the security needs of CMS and the techniques much easily
to deal with them.
Some of the important work related with this study is 3. htaccess and configuration file backup and
as follows: restore
A web site protection oriented remote backup and Configurations files which are useful for
recovery method : running website should be protected from attacks.
He Qian, Guo Yafeng, Wang Yong, in his thesis It is main file which provides security to other
describes that how we can take incremental abd CMS modules.
decremental backups of the website which will be used to
recover site during disaster. [1]. 4. Functionality to blacklist and whitelist
Website Regional Rapid Security Detection It is used to blacklist and whitelist IP
Method: addresses of the web surfers. It is recommended
Yong Fang ; Liang Liu, suggested that distributed to identify the search engine boat and spam boats
design, website regional rapid security detection method
can conduct security detection for the whole region by
adding detection module dynamically as needed and record
4. ANALYZE AND SUGGEST TOOL
results of detection. [2]
Research and solution of existing security Analyze and suggest tool is used to scan the CMS
problems in current internet website system : website for checking out possible threat inside the system.
It then analyze the website and generate the security
www.ijcat.com 810
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 809 - 811, 2014, ISSN:- 2319–8656
reports and suggest out some possible solutions and also Pritesh A.Taral received the B.E. degree in Computer Engineering
from S.A.O.E Pune, INDIA in 2011 and perusing M.E. degree in
provides option to incorporate them into the current CMS Computer Engineering from S.A.O.E , Pune
system
Prof. Balasaheb B.Gite is working as Head of the department of
Computer engineering at SAOE Pune India. He received the B.E.
degree in Computer Engineering from P.R.E.C Loni INDIA and
M.E. degree from W.C.E, Pune.
5. CONCLUSION
6. REFERENCES
[1] He Qian, Guo Yafeng, Wang Yong, Qiang Baohua1 “A
web site protection oriented remote backup and recovery
method” INSPEC Accession Number : 14022497 2014 IEEE
[2] Yong Fang ; Liang Liu, “Website Regional Rapid
Security Detection Method” 978-1-4799-0587-4 20 13 IEEE
[3] Gaoqi Wei, “Research and solution of existing security
problems in current internet website system”, 978-1-4244-
2584-6 20 08 IEEE
[4] Wenping Liu ; Xiaoying Wang ; Li Jin, “Design and
implementation of a website security monitoring system from
users' perspective”, 978-0-7695-4817-3/12 2012 IEEE
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 812 - 814, 2014, ISSN:- 2319–8656
Abstract : This paper deals with Sentence Validation - a sub-field of Natural Language Processing. It finds various applications in
different areas as it deals with understanding the natural language (English in most cases) and manipulating it. So the effort is on
understanding and extracting important information delivered to the computer and make possible efficient human computer
interaction. Sentence Validation is approached in two ways - by Statistical approach and Semantic approach. In both approaches
database is trained with the help of sample sentences of Brown corpus of NLTK. The statistical approach uses trigram technique based
on N-gram Markov Model and modified Kneser-Ney Smoothing to handle zero probabilities. As another testing on statistical basis,
tagging and chunking of the sentences having named entities is carried out using pre-defined grammar rules and semantic tree parsing,
and chunked off sentences are fed into another database, upon which testing is carried out. Finally, semantic analysis is carried out by
extracting entity relation pairs which are then tested. After the results of all three approaches is compiled, graphs are plotted and
variations are studied. Hence, a comparison of three different models is calculated and formulated. Graphs pertaining to the
probabilities of the three approaches are plotted, which clearly demarcate them and throw light on the findings of the project.
www.ijcat.com 812
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 812 - 814, 2014, ISSN:- 2319–8656
extract the E-R pairs in this sentence and match them from
database entries to calculate probability for semantic validity.
The same corpus data and test data for the above three
approaches are taken for comparison purposes. Graphs
pertaining to the results are plotted and major differences and
improvements are seen which are later illustrated and
analyzed.
Figure. 1 Parse Tree Generated by CFG Figure. 3 Testing results for Pure Statistical Markov Model
4. COMPLETE STRUCTURE
We have trained the database with 85% corpus and testing 4.2 Chunked-Off Markov Model
with the rest of 15% corpus we have. This has two advantages The second module is Chunked-Off Markov Model[3] -
- firstly we shall use the same ratio in all other approaches so training the database with corpus sentences in which all the
that we can compare them easily. Secondly it provides a nouns and named entities are replaced with their respective
threshold value for probability which will help us to type. This is implemented using the tagging and chunking
distinguish between correct and incorrect test sentences operations of NLTK. This solves the problem of Pure
depicting regions above and below threshold respectively. Statistical model that it is not able to deal with proper nouns.
Graphs are plotted between probability(exponential, in order For example, a corpus sentence has the trigram 'John eats pie'.
of 10) and length of the sentence(number of words). If a test sentence occurs like 'Mary eats pie', it will result in a
very low trigram probability. But if the trigram 'John eats pie'
is modified to 'PERSON eats pie', it will result in a better
comparison.
www.ijcat.com 813
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 812 - 814, 2014, ISSN:- 2319–8656
www.ijcat.com 814
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 815 - 821, 2014, ISSN:- 2319–8656
Abstract: Sonar imaging is one of the simplest technique for detection of under water drowned bodies. There is a need for design of
conventional beamforming which are robust and simple. Adaptive beamformer is used to improve the quality of the sonar image. As a
result we get an image containing more useful and correct information. The CORDIC computing technique a highly efficient method
to compute elementary functions like sine, cosine, translate, rotate values using CORDIC algorithm. The system simulation was
carried out using ModelSim and Xilinx ISE Design Suite 9.2i.. Matlab code is used to implement sin and cos using cordic angles and
amplitude response of beamformed data by optimized method in order to enlarge the validity region of beamforming. Synthesis
results of cordic shows the reduced memory requirement and less power consumption.
2. BACKGROUND THEORY
2.1.Beamforming
Beamforming is a type of signal processing technique used in
sensor arrays for directional signal transmission or reception.
Here the elements are combined in such a way that signals at
particular angles experience constructive interference while
others experience destructive interference[1]. Beamformers
are classified as either data independent or statistically
optimum, depending on how the weights are chosen. The Fig.1 Passive and Acive Sonar System
weights in a data independent beamformer do not depend on
the array data and are chosen to present a specified response needed during the signal processing system as
for all signal and interference scenarios. The weights in a shown in Fig.1
statistically optimum beamformer are chosen based on the
statistics of the array data to optimize the array response. The 2.4.Cordic Theory
statistics of the array data are not usually known and may Coordinate Rotational Digital Computer (CORDIC) is a set of
change over time so adaptive algorithms are typically used to shift-add algorithm known for computing a wide range
determine the weights. The adaptive algorithm is designed so trigonometric functions, hyperbolic, linear and logarithmic
the beamformer response converges to a statistically optimum functions also like multiplication division, data type
solution [6]. conversion, square root. It is highly efficient, low complexity.
The weights in a data independent beam former are The CORDIC algorithm has found in various applications
designed so that the beamformer response approximates a such as pocket calculator, numerical co-processors to high
desired response independent of the array data or data performers Radar signal processing, supersonic bomber.
statistics. This design objective is same as that for a classical Vector rotation can also be used for polar to rectangular and
FIR filter design. The simple delay and sum beam former is rectangular to polar conversions, for vector magnitude, and as
an example of the data independent beamforming. a building block in certain transforms such as the DFT and
In statistically optimum beam former the weighs DCT. The CORDIC algorithm provides an iterative method
are chosen based on the statistics of the data received at the of performing vector rotations by arbitrary angles using only
array. The goal is to optimize the beam former response so shifts and adds. The algorithm, credited to Volder[4], is
that the output signal contains minimal contributions due to derived from the general (Givens) rotation transform:
the noise and signals arriving from directions other than the x=׳x.cos (ɸ) - y.sin (ɸ) -------------(1)
desired direction. The Frost beamformer is a statistically
optimum beam former. Other statistically optimum y=׳x.cos (ɸ) + y.sin (ɸ) --------------(2)
beamformers are Multiple Side lobe Canceller and
Maximization of the signal to noise ratio. .
These can be rearranged so that:
2.2.Sonar Imaging
Sonar (an acronym for SOund Navigation and Ranging) is a
technique that uses sound propagation (usually underwater, as
in submarine navigation) to navigate, communicate with or
detect objects on or under the surface of the water, such as
other vessels. Two types of technology share the name
"sonar": passive sonar is essentially listening for the sound
made by vessels; active sonar is emitting pulses of sounds and
listening for echoes. Sonar may be used as a means of
acoustic location and of measurement of the echo
characteristics of "targets" in the water. Acoustic location in
air was used before the introduction of radar. Sonar may also
be used in air for robot navigation, and SODAR (upward
looking in-air sonar) is used for atmospheric investigations. Fig 2. Rotation of sin and cos
The term sonar is also used for the equipment used to generate
and receive the sound. x=׳cos (ɸ). [x – y.tan (ɸ)] --------------(3)
2.3.Active and Passive Sonar System
y=׳cos (ɸ). [x + y.tan (ɸ)] ---------------(4)
Active sonar or passive sonar, when receiving the acoustic
signal reflected from the target, the information included in
the signal cannot be directly collected and used without
technical signal processing. To extract the efficient and useful
. information’s from the mixed signal, some steps should be
taken to transfer sonar data from raw acoustic data reception
to detection output.
www.ijcat.com 816
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 815 - 821, 2014, ISSN:- 2319–8656
Fig 3. Input and output of rotation for rotation mode duplicating each of the three difference equations in hardware
as shown in Figure 1. The decision function, di, is driven by
So far, nothing is simplified. However, if the rotation angles the sign of the y or z register depending on whether it is
are restricted so that tan()2-i, the multiplication by the tangent operated in rotation or vectoring mode. In operation, the
term is reduced to simple shift operation. Arbitrary angles of initial values are loaded via multiplexers into the x, y and z
rotation are obtainable by performing a series of successively registers. Then on each of the next n clock cycles, the values
smaller elementary rotations. If the decision at each iteration, from the registers are passed through the shifters and adder-
i, is which direction to rotate rather than whether or not to subtractors and the results placed back in the registers. The
rotate, then the cos(i) term becomes a constant (because shifters are modified on each iteration to cause the desired
cos(i) = cos(-i)). The iterative rotation can now be expressed as: shift for the iteration. Likewise, the ROM address is
incremented on each iteration so that the appropriate
Xi+1 =Ki[ xi – di.yi.2-i ] ---------(5) elementary angle value is presented to the z adder-subtractor.
On the last iteration, the results are read directly from the
Yi+1 =Ki[ xi + di.yi.2-i] --------(6) adder-subtractors. Obviously, a simple state machine is
required keep track of the current iteration, and to select the
Where degree of shift and ROM address for each iteration. The design
depicted in Figure 1 uses word-wide data paths (called bit-
Ki= cos(tan-12-i)= 1/sqrt(1+ 2-2i) ----(7) parallel design). The bit-parallel variable shift shifters do not
map well to FPGA architectures because of the high fan-in
di= -----------------------------------(8) required. If implemented, those shifters will typically require
several layers of logic (i.e., the signal will need to pass
Removing the scale constant from the iterative equations through a number of FPGA cells). The result is a slow design
yields a shift-add algorithm or vector rotation. The product that uses a large number of logic cells.
of the Ki's can be applied elsewhere in the system or treated
as part of a system processing gain. That product approaches 3. PROPOSED WORK
0.6073 as the number of iterations goes to i infinity. Digital input pulse is passed to find the angle or detection of
Therefore, the rotation algorithm has a gain, An of object under water. In Fig.3.Beamforming can be used at both
approximately 1.647. The exact gain depends on the number the transmitting and receiving ends in order to achieve spatial
of iterations, and obeys the relation selectivity, the data is transmitted to underwater sonar system,
sonar is used to detect the underwater objects and finds the
An=∏ sqrt(1+ 2-2i) -------------------(9) angle elevation. The beamformed data is transmitted; at the
receiver end beam formation data is generated. The generated
The angle of a composite rotation is uniquely defined by the beam formation data will be having interference and noise
sequence of the directions of the elementary rotations. That error that will be reduced by using optimization technique.
sequence can be represented by a decision vector. The set of Optimized cordic beamforming will eliminate all the
all possible decision vectors is an angular measurement interference which generated at receiver end. Final
system based on binary arctangents. Conversions between this optimization beamforming data is obtained
angular system and any other can be accomplished using
look-up. A better conversion method uses an additional
adder-subtractor that accumulates the elementary rotation Digital Beam Data
angles at each iteration. The elementary angles can be
expressed in any convenient angular unit. Those angular values Input forma- transmitti
are supplied by a small lookup table (one entry per iteration) Pulse tion ng to
or are hardwired, depending on the implementation. The
angle accumulator adds a third difference equation to the from under
algorithm. transmit water
Zi+1 = Zi + di(tan -1 2-i )----------------(10) ter sonar
system
Obviously, in cases where the angle is useful in the
arctangent base, this extra element is not needed. The CORDIC
rotator is normally operated in one of two modes. The first,
called rotation by Volder[4], rotates the input vector by a
specified angle (given as an argument). The second mode,
called vectoring, rotates the input vector to the x axis
Output Optimiz Receiver
2.5. Implementation in an FPGA
There are a number of ways to implement a CORDIC Beam ed (Beam
processor. The ideal architecture depends on the speed versus Data Cordic(B formation
area tradeoffs in the intended application. First we will
examine an iterative architecture that is a direct translation eam Data
from the CORDIC equations. From there, we will look at a forming) Generate)
minimum hardware solution and a maximum performance
solution.
2.6.Iterative CORDIC Processors
An iterative CORDIC architecture can be obtained simply by Fig 4.. Beamforming for underwater sonar
www.ijcat.com 817
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 815 - 821, 2014, ISSN:- 2319–8656
3.1.Program Flow Chart: Fig 5. shows the flow chart in which initially the
CORDIC values are sampled, the antennas are used to detect
the angles and object of beam form underwater sonar with
start sampled bit rate. The detected angles are taken as input data
where the beam data is formed, the obtained beam data are
sampled according to mathematical calculations under
CORDIC algorithm, the obtained beam data samples are
computed as quad phase and In phase. The received beam
Initialize cordic data contains noise and interference which are reduced and
values eliminated using underwater noise model. The beam data is
computed for each antenna and its angles, the error beam data
is finally computed to obtain noiseless beam data. The
obtained output is in the form of optimized beam form data.
Bit Rate No.of
3.2. Architecture:
Anten The architecture is shown in Fig.6 in which input signals are
given to memory. The memory is used to store the data of
input signals. The signals are transmitted to detect the target
or object in underwater beam form data. Once the target is
detected and beam form data is generated. The received beam
Data sample
form data is up sampled and Down sampled. The adder is
rate used to combine the images received and stored in memory.
The generated beam form data signals from sonar are given to
Input data CORDIC algorithm
Beam
Formation
Compute quad
and in phase
Beam Data
Received
Generator
underwater
Fig.6. Data path algorithm
www.ijcat.com 818
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 815 - 821, 2014, ISSN:- 2319–8656
Y’=0 ----------------(15)
Ө’=atan(x/y) ----------------(16)
www.ijcat.com 819
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 815 - 821, 2014, ISSN:- 2319–8656
Fig.11. Phase Graph of Input Data Fig.14.shows the loss of data which is less compared to direct
method.
Fig.11. shows the input pulse sent from the transmitter side to
underwater to detect the target.
www.ijcat.com 820
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 815 - 821, 2014, ISSN:- 2319–8656
7. REFERENCES
[1] V. Murino and A.Trucco , “Three-dimensional image
generation and processing in under acoustic vision,”
vol88,n0.12 dec 2000
[2] A.Davis and A.Lugsdin, “High speed underwater
inspection for port and harbour security using coda
Echoscope 3D sonar,” 2005,pp.2006-2011 .
[3] R.K.Hansen and P.A Andersen, “The application of real
time 3D acoustical imaging,” OCEANS1998 pp.738-741
[4] M. Palmese and A. Trucco, “Digital near field
beamforming for efficient 3-D underwater acoustic
image generation,” in Proc. IEEE Int. Workshop Imaging
Syst. Tech., 2007, pp. 1-5.
[5] M. Palmese and A. Trucco, “From 3-D sonar images to
augmented reality models for objects buried on the
seafloor,” IEEE Trans. Instrum. Meas., vol. 57, no. 4, pp.
820-828, Apr. 2008.
www.ijcat.com 821
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 822 - 826, 2014, ISSN:- 2319–8656
Abstract: Ethernet a popular choice for metropolitan-area networks (MAN) due to simplicity, cost effectiveness and scalability. The
Spanning-Tree based switching mechanism, which is considered to be very efficient at avoiding switching loops in LAN environment,
is a performance bottleneck in Metro network context. Handling of link failure is an important issue in metro Ethernet networks. A
link failure may result in serious service disruptions. A local restoration method for metro Ethernet with multiple spanning trees, which
aims at fast handling of single link failures in a distributed manner, have been proposed in the literature. In this paper, we propose a
local restoration mechanism that uses MULTILINK algorithm for solving multiple link failures
Keywords: Metropolitan Area Networks (MAN), Ethernet, Spanning Tree Protocol, RSTP.
www.ijcat.com 822
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 822 - 826, 2014, ISSN:- 2319–8656
www.ijcat.com 823
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 822 - 826, 2014, ISSN:- 2319–8656
www.ijcat.com 824
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 822 - 826, 2014, ISSN:- 2319–8656
Else
Route the packet through RLink.
Fig.2 Algorithm Y
IF TLink
is up?
N
IF Blink
is
down?
N Y
IF
NB=1?
Stop
www.ijcat.com 825
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 822 - 826, 2014, ISSN:- 2319–8656
6. CONCLUSION
The existing system only handles single link failures in the
metro Ethernet. The proposed system is an enhancement of
the existing system. This system defines a possible approach
to handle these multiple link failures. It also uses the concept
of local restoration mechanism provided in the existing
system in its restoration module. The restoration module uses
an algorithm namely MULTILINK. Two bits in the Ethernet
IP packet header are used in this algorithm. The occurrence of
multiple link failure is a rare event. When implemented
properly, the proposed system solves the problem of multiple
link failures in metro Ethernet network.
7. REFERENCES
[1] Jian Qiu, Mohan Gurusamy, ―Local Restoration With
Multiple Spanning Trees in Metro Ethernet Networks‖,
IEEE/ACM Transactions On Networking, Vol. 19, No. 2,
April 2011
[2] A. Meddeb, ―Why Ethernet WAN transport,‖ IEEE
Commun. Mag.,vol. 43, no. 11, pp. 136–141, Nov. 2005.
[3] C. Antal, L. Westberg, A. Paradisi, T. R. Tronco, and V.
G. Oliveira, ―Fast failure handling in Ethernet network,‖
in Proc IEEE ICC, 2006, vol. 2, pp. 841–846.
[4] Standard for Local and Metropolitan Area Networks—
Media Access Control (MAC) Bridges, IEEE 802.1d,
1998.
[5] Standard for Local and Metropolitan Area Networks—
Rapid Reconfiguration of Spanning Tree, IEEE 802.1w,
2001.
[6] Standard for Local and Metropolitan Area Networks—
Virtual Bridged Local Area Networks, IEEE 802.1q,
1999.
[7] J. Farkas, C. Antal, G. Toth, and L.Westberg,
―Distributed resilient architecture for Ethernet networks,‖
in Proc. DRCN, 2005, pp. 515–522.
[8] K. Goplan, S. Nanda, and T. Chiueh, ―Viking: A
multiple-spanning tree Ethernet architecture for
metropolitan area and cluster networks,‖ in Proc. IEEE
INFOCOM, 2004, pp. 2283–2294.
www.ijcat.com 826
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 827 - 830, 2014, ISSN:- 2319–8656
Abstract: Popularity of WWW increasing day by day which results in the increase of web based services , due to which web is now
largest data repository. In order to handle this incremental nature of data various prediction techniques are used. If the prefetched pages
are not visited by user in their subsequent access there will be wastage network bandwidth as it is in limited amount. So there is critical
requirement of accurate prediction method. As the data present on web is heterogeneous in nature and incremental in nature, during the
pre-processing step hierarchical clustering technique is used. Then using Markov model category and page prediction is done and
lastly page filtering is done using keywords.
Keywords: formatting Hierarchal clustering; markov model; page prediction; category prediction ;
www.ijcat.com 827
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 827 - 830, 2014, ISSN:- 2319–8656
Preprocessing
Cluster view
Hierarchical clustering
(1)
www.ijcat.com Clusetr1 Cluster 828
2
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 827 - 830, 2014, ISSN:- 2319–8656
(6)
Where
(7)
4. CONCLUSION
Figure 2. Sample web sessions with corresponding 1st and 2nd As there is large amount of data on web pages on many
order transition probability matrices [7]. websites, So it is better to place them according to their
category. In this paper users browsing behavior is firstly
preprocessed using hierarchical clustering then prediction is
3.4 SIMILARITY MATRIX done in three phases. In first phase category prediction is done
using Markov model then in second phase page prediction is
Similarity between any two user user i and user j, can be done. And lastly keyword based filtering is done which gives
calculated using Euclidean distance given in equation (3) . more accurate results.
5. REFERENCES
.
www.ijcat.com 829
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 827 - 830, 2014, ISSN:- 2319–8656
[7] Deshpande M and Karypis G (2004), “Selective Markov [10] A. Anitha, “A New Web Usage Mining Approach for
Models for Predicting Web-Page Accesses”, ACM Next Page Access Prediction”, International Journal of
Transactions on Internet Technology Computer Applications, Volume8–No.11,October2010.
(TIOIT),Vol.4,No.2,pp.163-184.
[11] Mehrdad Jalali, Norwati Mustapha, Md. Nasir Sulaiman,
[8] UCI KDD archive, http://kdd.ics.uci.edu/ Ali Mamat, “WebPUM: A Web-Based Recommendation
System to Predict User Future Movements” Expert
[9] V.V.R. Maheswara Rao, Dr. V. Valli Kumari” An Systems with Applications 37 , 2010.
Efficient Hybrid Predictive Model to Analyze the
Visiting Characteristics of Web User using Web Usage [12] www.microarrays.ca/services/hierarchical_clustering.pdf
Mining” 2010 International Conference on Advances in
Recent Technologies in Communication and Computing
IEEE.
www.ijcat.com 830
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 831 - 835, 2014, ISSN:- 2319–8656
Abstract: Performance prediction and evaluation for SOA based applications assist software consumers to estimate their applications
based on service specifications created by service developers. Incorporating traditional performance models such as Stochastic Petri
Nets, Queuing Networks, and Simulation present drawbacks of SOA based applications due to special characteristics of SOA such as
lose coupling, self-contained and interoperability. Although, researchers have suggested many methods in this area during last decade,
none of them has obtained popular industrial use. Based on this, we have conducted a comprehensive survey on these methods to
estimate their applicability. This survey classified these approaches according to their performance metrics analyzed, performance
models used, and applicable project stage. Our survey helps SOA architects to select the appropriate approach based on target
performance metric and researchers to identify the SOA state-of-art performance prediction.
www.ijcat.com 831
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 831 - 835, 2014, ISSN:- 2319–8656
based on specifications defined in WSDL, step 6 the response o The potential problems of XML which is the
generated by the provider is also in tag based XML format. standard message format increases the time needed
to process a request.
o The time needed to discover the services through
the directory either in design time or run time.
o Rules that govern services contain a business
process by business process’s need.
Directory o Adaptation of service composition by adding new
service or adapting existing services.
WSDL Service Description o Think time is an elapsed time between the end of a
response time generated by a service and the
(Find)
(Register)
(Locate) beginning of an end user’s request [4].
(Publish)
4.2 Throughput
SOAP SOAP Throughput defined as the number of requests SOA
application can process at a given period of time. There are
(Bind) two metrics for throughput; throughput of a service and
throughput of a business process [4] as Figure 3 stated.
(XML) Service Request
Service Service The value range of these two metrics service throughput and
Provider Consumer business process throughput must be greater than zero. The
(XML) Service Response
higher the values indicate a better SOA application
(Execute) performance.
4. SERVICE-ORIENTED
ARCHITECTURE PERFORMANCE
METRICS
4.1 Service Response Time
Service Response Time is the measure of the time between the
end of a request to a service and the beginning of the time
service provider response. There are many considerations to
measure service response time [4] as Figure 2 stated. The
main reasons that cause low performance of SOA based
applications are:
www.ijcat.com 832
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 831 - 835, 2014, ISSN:- 2319–8656
Business
Process
Request
Message
Processing
Time
Business
Process
Execution
Waiting Time Service Throughput
Throughput
Service
Discovery Business Process
Time Throughput
Service
Adaptation
Time
Business
Process
Transmissio
n Time
Service
Composition Figure. 3 Sub-metrics of SOA Throughput
Service
Time Request
Message
Processing
Time
Service Business
Process Business
Response CPU Usage
Processing Process
Time Logic Service
Time
Execution Execuation
Time Waiting Time
Input/Output
Service Activity
Business Request Service Logic
Process Transmission Execuation Time
Response Time
Time
Secondary Communication
Service Storage Requst Resorce Utlization Devices
Processing Transmission
Time Time
Secondary
Number of
Stoarge Response
Database Calls
Transmission
Business
Time
Process
Response
Message
Processing Service
Time Response
Message Figure. 4 Sub-metrics of SOA Resource Utilization
Processing Time
www.ijcat.com 833
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 831 - 835, 2014, ISSN:- 2319–8656
Author Name/ Main Objective Approach used Metrics Analyzed Performance Method’s Applicable
Year Model Validation Project Stage
Kounev, Designing systems Use dynamic Response time and Queuing Petri net Compared with Runtime
Samuel, et al. with build-in self- architecture-level Resource Model PCM model result
[6], 2010 aware performance performance utilization
and resource model at run-time
management for online
capabilities performance and
resource
management
Liu, et al. [7], Develop a Measure primitive Throughput and Queuing Network Compared with the Runtime
2007 performance performance response time Model results of Microsoft
model for overheads of Web Stress Tool
predicting runtime service routing
performance based activities in the
on COTS ESB ESB
(Enterprise Service
Bus)
Tribastone, et al. Present a method Modeling the Response time, Layered Queuing Compared with Design time
[8], 2010 for performance system using Processor Network Mobile Payment
predication of UML and two utilization Model case study
SOA at early stage profiles, performance result
of development. UML4SOA, and
MARTE
Teixeira, et al. Propose approach The model uses Resource Stochastic Petri Compared with Design time
[9], 2009 to estimate Petri Net consumption, Nets (Rud et al)
performance of formalism to Service levels Model Analytical
SOA represent the degradation Method and values
process and from real
estimate its applications
performance
using simulation.
Punitha, et al. Developing an Building and Response time, Queuing Network Prototype SOA Design time
[11], 2008 architectural measuring the throughput, load Model Application has
performance performance capacity, heavily been implemented
model for SOA model using loaded and measured
ACME language components.
Brüseke, et al. Developing Comparing the Response time Palladio Applied on two case Design time
[12], 2014 PBlaman observed response Component Model studies
(Performance time of each (PCM)
Blame Analysis ) component in a
failed test case to
expected response
time from the
contract
Reddy, et al. Modeling Web Simulate the Response time and SMTQA Model Applied on case Design time
[13], 2011 Service using model using Server utilization study
UML Simulation of
Multi-tiered
Queuing
Applications
(SMTQA)
www.ijcat.com 834
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 831 - 835, 2014, ISSN:- 2319–8656
Marzolla, et al. Present a multi- Approach for Response time and Queuing Network Prototype tool called Both Design
[14], view approach for performance throughput Model bpe12qnbound time and Run
2007 performance assessment of time
prediction of SOA Web Service
based applications workflows
for users and described using
providers annotated BPEL
and WSDL
specification
[6] Kounev, S., Brosig, F., Huber, N., & Reussner, R. (2010,
6. CONCLUSION July). Towards self-aware performance and resource
We have surveyed the state-of-art in the research of management in modern service-oriented systems.
performance prediction methods for service-oriented In Services Computing (SCC), 2010 IEEE International
architecture based applications. The survey categorized the Conference on(pp. 621-624). IEEE.
approaches according to the performance metrics analyzed,
performance model, method validation, and approach [7] Liu, Y., Gorton, I., & Zhu, L. (2007, July). Performance
applicable stage. prediction of service-oriented applications based on an
enterprise service bus. In Computer Software and
The field of performance evaluation and prediction for Applications Conference, 2007. COMPSAC 2007. 31st
service-oriented architecture based application has been Annual International(Vol. 1, pp. 327-334). IEEE.
developed and matured over the last decade. Many tools and
ideas have been implemented as good software engineering [8] Tribastone, M., Mayer, P., & Wirsing, M. (2010).
practice and should lead the creation of new approaches. Performance prediction of service-oriented systems with
layered queueing networks. In Leveraging Applications
Our survey helps both architects and researchers. Architects of Formal Methods, Verification, and Validation (pp. 51-
can obtain a complete view of the performance evaluation and 65). Springer Berlin Heidelberg.
prediction approaches proposed to transfer them to industry,
on the other hand researchers can align themselves with the [9] Teixeira, M., Lima, R., Oliveira, C., & Maciel, P. (2009,
proposed approaches and add more features in the future to October). Performance evaluation of service-oriented
enhance and enrich the area. architecture through stochastic Petri nets. InSystems,
Man and Cybernetics, 2009. SMC 2009. IEEE
7. REFERENCES International Conference on (pp. 2831-2836). IEEE.
[1] Bianco, P., Kotermanski, R., & Merson, P. F. (2007).
Evaluating a service-oriented architecture. [10] Balsamo, S., Mamprin, R., & Marzolla, M. (2004).
Performance evaluation of software architectures with
[2] Krafzig, D., Banke, K., & Slama, D. (2005). Enterprise queuing network models. Proc. ESMc, 4.
SOA: service-oriented architecture best practices.
Prentice Hall Professional. [11] Punitha, S., & Babu, C. (2008, September). Performance
prediction model for service oriented applications.
[3] Erl, T. (2004). Service-Oriented Architecture. Concepts, In High Performance Computing and Communications,
Technology, and Design. Tavel, P. 2007 Modeling and 2008. HPCC'08. 10th IEEE International Conference
Simulation Design. AK Peters Ltd. on (pp. 995-1000). IEEE.
[4] Her, J. S., Choi, S. W., Oh, S. H., & Kim, S. D. (2007, [12] Brüseke, F., Wachsmuth, H., Engels, G., & Becker, S.
October). A framework for measuring performance in (2014). PBlaman: performance blame analysis based on
service-oriented architecture. In Next Generation Web Palladio contracts. Concurrency and Computation:
Services Practices, 2007. NWeSP 2007. Third Practice and Experience.
International Conference on (pp. 55-60). IEEE.
[13] Reddy, C. R. M., Geetha, D. E., Srinivasa, K. G., Kumar,
[5] Abowd, G., Bass, L., Clements, P., Kazman, R., & T. S., & Kanth, K. R. (2011). Predicting performance of
Northrop, L. (1997).Recommended Best Industrial web services using SMTQA. International Journal of
Practice for Software Architecture Evaluation (No. Computer Science Information Technology, 1(2), 58-66.
CMU/SEI-96-TR-025). CARNEGIE-MELLON UNIV
PITTSBURGH PA SOFTWARE ENGINEERING [14] Marzolla, M., & Mirandola, R. (2007). Performance
INST. prediction of web service workflows. In Software
Architectures, Components, and Applications (pp. 127-
144). Springer Berlin Heidelberg.
www.ijcat.com 835
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 836 - 838, 2014, ISSN:- 2319–8656
Abstract Transmitting the geo-location information of a target via wireless networks is effective when both the target and the
tracker are within Wi-Fi coverage area; the 802.11 wireless networks are not always accessible. When the target or the tracker is
unable to access Wi-Fi, it is impossible to perform location tracking. Therefore, SMS is a relatively more reliable and flexible
solution because of its widespread use. In this system, a device is equipped with a global system for mobile communications
(GSM) modem and a GPS unit. It transmits short messages containing its GPS coordinates to the server at 30-s intervals.
Although transmitting the geo-location information of a target via wireless networks is effective when both the target and the
tracker are within Wi-Fi coverage area, the 802.11 wireless networks are not always accessible. When the target or the tracker is
unable to access Wi-Fi, it is impossible to perform location tracking. In this System, a novel method called location-based
delivery (LBD), which combines the short message service (SMS) and global position system (GPS). LBD reduces the number of
short message transmissions while maintaining the location tracking accuracy within the acceptable range. The proposed
approach, LBD, consists of three primary features: Short message format, location prediction, and dynamic threshold. The
defined short message format is proprietary.
Key Words: Short Message Service (SMS), Location Tracking, Mobile Phones, Prediction Algorithms, Global Positioning
System (GPS).
www.ijcat.com 836
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 836 - 838, 2014, ISSN:- 2319–8656
threshold, the target transmitsa shortmessage to the tracker 3. LOCATION BASED DELIVERY
to update its current location. Thedynamic threshold Location based services (LBS) a novel class of computer
maintains the location tracking accuracy andnumber of application, which combines the short message services
short messages on the basis of the moving speed ofthe (SMS) and global position system (GPS). As such LBS is
target.The simulations performed to test the performance of an informative service and number of uses in social
LBD show that compared with other related works; the networking today as an entertainment service, which is
proposed LBD minimizes the number of short message accessible with mobile devices through the mobile network
transmissions while maintaining the location prediction and which uses information on the geographical position of
accuracy within the acceptable range. the mobile device.
www.ijcat.com 837
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 836 - 838, 2014, ISSN:- 2319–8656
5. REFERENCES
[1] H. H. Lee, I. K. Park, and K. S. Hong, “Design and
implementation ofa mobile devices-based real-time
location tracking,” in Proc. UBICOMM, 2008, pp. 178–
183.
[2] Z. Tian, J. Yang, and J. Zhang, “Location-based
services applied to aelectric wheelchair based on the GPS
and GSM networks,” in Proc. ISA,2009, pp. 1–4.
[3] I. Lita, I. B. Cioc, and D. A. Visan, “A new approach of
automobile localizationsystem using GPS and GSM/GPRS
transmission,” in Proc. ISSE,2006, pp. 115–119.
[4] P. Perugu, “An innovative method using GPS tracking,
WINS technologiesfor border security and tracking of
vehicles,” in Proc. RSTSCC, 2010, pp. 130–133.
[5] S. A. Hameed, O. Khalifa, M. Ershad, F. Zahudi, B.
Sheyaa, and W.Asender, “Car monitoring, alerting, and
tracking model: Enhancementwith mobility and database
facilities,” in Proc. ICCCE, 2010, pp. 1–5.
[6] R. E. Anderson, A. Poon, C. Lustig, W. Brunette, G.
Borriello, and B. E.Kolko, “Building a transportation
information system using only GPS andbasic SMS
infrastructure,” in Proc. ICTD, 2009, pp. 233–242.
[7] W. J. Choi and S. Tekinay, “Location-based services for
next-generationwireless mobile networks,” in Proc. IEEE
VTC, 2003, pp. 1988–1992.
[8] R. E. Anderson, W. Brunette, E. Johnson, C. Lustig, A.
Poon, C. Putnam,
O. Salihbaeva, B. E. Kolko, and G. Borrielllo,
“Experiences with atransportation information system that
uses only GPS and SMS,” in Proc.ICTD, 2010.
[9] A. Civilis, C. S. Jensen, and S. Pakalnis, “Techniques
for efficient roadnetwork-based tracking of moving
objects,” IEEE Trans. Knowl.DataEng., vol. 17, no. 5, pp.
698–712, 2005.
[10] M. Zahaby, P. Gaonjur, and S. Farajian, “Location
tracking in GPS usingKalman filter through SMS,” in Proc.
IEEE EUROCON, 2009, pp. 1707–1711.
[11] A. Civilis, C. S. Jensen, J. Nenortaite, and S. Pakalnis,
“Efficient trackingof moving objects with precision
guarantees,” in Proc. MOBIQUITOUS,2004, pp. 164–173.
[12] Y. Y. Xiao, H. Zhang, and H. Y. Wang, “Location
prediction for trackingmoving objects based on grey
theory,” in Proc. FSKD, 2007, pp. 390–394.
[13] P. H. Tseng, K. T. Feng, Y. C. Lin, and C. L. Chen,
“Wireless locationtracking algorithms for environments
with insufficient signal sources,”IEEE Trans. Mobile
Comput., vol. 8, no. 12, pp. 1676–1689, 2009.
[14] R. Bajaj, S. L. Ranaweera, and D. P. Agrawal, “GPS:
Location-trackingtechnology,” Computer, vol. 35, no. 4,
pp. 92–94, 2002.
[15] Movable Type Scripts. (2012 June).[Online].
Available: http://www.mova
ble-type.co.uk/scripts/latlong.html
www.ijcat.com 838
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 839 - 843, 2014, ISSN:- 2319–8656
Abstract:
Spam is defined as redundant and unwanted electronica letters, and nowadays, it has created many problems in business life such as
occupying networks bandwidth and the space of user’s mailbox. Due to these problems, much research has been carried out in this
regard by using classification technique. The resent research show that feature selection can have positive effect on the efficiency of
machine learning algorithm. Most algorithms try to present a data model depending on certain detection of small set of features.
Unrelated features in the process of making model result in weak estimation and more computations. In this research it has been tried
to evaluate spam detection in legal electronica letters, and their effect on several Machin learning algorithms through presenting a
feature selection method based on genetic algorithm. Bayesian network and KNN classifiers have been taken into account in
classification phase and spam base dataset is used.
www.ijcat.com 839
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 839 - 843, 2014, ISSN:- 2319–8656
more efficient and useful features decrease time complexity In this research, we used weighted F-score for calculate
and increase accuracy [17]. the fitness value of each chromosome. The algorithm
starts by randomly initializing a population of N
3. GENETIC ALGORITHMS number of initial chromosome.
A genetic algorithm (GA) is one of a number of heuristic
techniques that are based on natural selection from the 4.1.2 Cross over
population members attempt to find high-quality solutions to The crossover is the most important operation in GA.
large and complex optimization problems. This algorithm can Crossover as name suggests is a process of
identify and exploit regularities in the environment, and recombination of bit strings via an exchange of
converges on solutions (can also be regarded as locating the segments between pairs of chromosomes. There are
local maxima) that were globally optimal [18]. This method is various kinds of crossover. In one point cross-over, a bit
very effective and widely used to find-out optimal or near position is randomly selected that need to change. In
optimal solutions to a wide variety of problems. The genetic this process, a random number is generated which is a
algorithm repeatedly modifies a population of individual number (less than or equal to the chromosome length)
solutions. At each step the genetic algorithm tries to select the as the crossover position [21]. Here one crossover point
best individuals. From the current “parent” population genetic is selected, binary string from beginning of
algorithm creates “children”, who constitute next generation. chromosome to the crossover point is copied from one
Over successive generations the population evolves toward an parent, the rest is copied from the second parent[22].
optimal solution. The genetic algorithm uses three main rules 4.1.3. Proposed mutation
at each step to create the next generation. Select the Mutation has the effect of ensuring that all possible
individuals, called parents that contribute to the population at chromosomes can maintain good gene in the newly
the next generation. Crossover rules that combine two parents generated chromosomes. In our approach, Mutation
to form children for the next generation. Mutation rules, apply operator is a two-steps process, and is a combination of
random changes to individual parents to form children. random and substitution mutation operator. Also is
occurs on the basis of two various mutation rates.
4. FEATURE SELECTION Mutation operator firstly events substitution step with
Features selection approaches are usually employed to the probability of 0.03. In each generation, the best
reduce the size of the feature set, and to select a subset chromosome involving better features and higher fitness
of the original features. We use the proposed genetic is selected, and it substitutes for the weakest
algorithms to optimize the features that contribute chromosome having lesser fitness than others. ( َایي هزحل
significantly to the classification. َباعث اًتقال کزّهْسّم بزتز ًسل جاری بَ ًسل بعذی هی ضْد ک
)ُوگزایی سزیع الگْریتن را ُن بَ دًبال خْاُذ داضتOtherwise, it
enters the second mutation step with probability of 0.02.
4.1. Feature Selection Using Proposed This step changes some gens of chromosome randomly
Genetic Algorithm by inverting their binary cells. In fact the second is
considered to prevent reducing exploration capability of
In this section, the method of feature selection by using search space to keep diversity in other chromosomes.
the proposed genetic Algorithm has been presented. The Generally mutation probability is equal to 0.05.
procedure of the proposed method has been stated in
details in the following section. 5. RESULTS SIMULATION
4.1.1. Initialize population In order to investigate the impact of our approach on
In the genetic algorithm, each solution to the feature email spam classification, spam base data set that
selection problem is a string of binary numbers, called downloaded from the UCI Machine Learning
chromosome. In this algorithm initial population is Repository are used [23]. Data set of Spam base
generated randomly. IN feature representation as a involving 4601 emails was proposed by Mark Hopkins,
chromosome, if the value of chromosome [i] is 1, the ith and his colleagues. In This data set that is divided into
feature is selected for classification, while if it is 0, then
two parts, 1 shows spam, and zero indicates non-spam.
these features will be removed [19,20]. Figure 1 shows
feature presentation as a chromosome. This data set involves 57 features with continuous
values. In simulation of the proposed method, training
Chromosome:
set involving 70% of the main data set and two
experimental sets have been separately considered for
...
feature selection and classification. Each one involves
15% of the main data set. After performing feature
1 0 1 ... 1 0 selection using the training set, the test set was used to
evaluate the selected subset of features. The evaluation
of the overall process was based on weighted f-score
Figure 1. Feature Subset: {F1,F3 , …, Fn-1 } which is a suitable measure for the spam classification
www.ijcat.com 840
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 839 - 843, 2014, ISSN:- 2319–8656
Initial population
80
Mutation rate1
0.03
Mutation rate2
0.02
Crossover
0.7
Generations
100
www.ijcat.com 841
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 839 - 843, 2014, ISSN:- 2319–8656
of selected features
all evaluation criteria have been considerably improved. Sahami, M., et. al.,(1998). A Bayesian Approach to
ّ (پس رّش پیطٌِادی تاثیز قابل هالحظَ ای بز کاُص ّیژگی ُا Filtering Junk E-Mail, In Learning for Text
Categorization, AAAI Technical Report, U.S.A.
).بِبْد دقت داضتَ است/افشایصWe can use of parameter
[8] Riley. J. (2002). An evolutionary approach to training
optimization in this work also the proposed algorithm
Feed-Forward and Recurrent Neural Networks", Master
can be combined with other classification algorithms in thesis of Applied Science in Information Technology,
the future. Department of Computer Science, Royal Melbourne
Institute of Technology, Australia.
REFERENCE [9] Clark, et. al. (2003). A Neural Network Based Approach
[1] GOWEDER, A. M., RASHED, T., ELBEKAIE, A., & to Automated E-Mail Classification, IEEE/WIC
ALHAMMI, H. A. (2008). An Anti-Spam System Using International Conference on Web Intelligence.
Artificial Neural Networks and Genetic Algorithms. [10] Branke, J. (1995). Evolutionary algorithms for neural
Paper presented at the Proceedings of the 2008 network design and training, In Proceedings 1st Nordic
International Arab Conference on Information Workshop on Genetic Algorithms and its Applications,
Technology. Finland.
[2] Bruening, P.(2004). Technological Responses to the [11] Yao. X., Liu. Y. (1997). A new evolutionary system for
Problem of Spam: Preserving Free Speech and Open evolving artificial neural networks", IEEE Transactions
Internet Values. First Conference on E-mail and Anti- on Neural Networks.
Spam.
[12] Wang, H.-b., Y. Yu, and Z. Liu. (2005) SVM classifier
[3] Graham, P.(2003). A Plan for Spam. MIT Conference on incorporating feature selection using GA for spam
Spam. detection, in Embedded and Ubiquitous Computing–
[4] William, S., et. al. (2005). A Unified Model of Spam EUC 2005., Springer. p. 1147-1154.
Filtration, MIT Spam Conference, Cambridge.
www.ijcat.com 842
International Journal of Computer Applications Technology and Research
Volume 3– Issue 12, 839 - 843, 2014, ISSN:- 2319–8656
www.ijcat.com 843