Cyber Security Compress
Cyber Security Compress
Cyber
Security
Proceedings of CSI 2015
Advances in Intelligent Systems and Computing
Volume 729
Series editor
Janusz Kacprzyk, Polish Academy of Sciences, Warsaw, Poland
e-mail: [email protected]
The series “Advances in Intelligent Systems and Computing” contains publications on theory,
applications, and design methods of Intelligent Systems and Intelligent Computing. Virtually all
disciplines such as engineering, natural sciences, computer and information science, ICT,
economics, business, e-commerce, environment, healthcare, life science are covered. The list of
topics spans all the areas of modern intelligent systems and computing such as: computational
intelligence, soft computing including neural networks, fuzzy systems, evolutionary computing and
the fusion of these paradigms, social intelligence, ambient intelligence, computational neuro-
science, artificial life, virtual worlds and society, cognitive science and systems, Perception and
Vision, DNA and immune based systems, self-organizing and adaptive systems, e-Learning and
teaching, human-centered and human-centric computing, recommender systems, intelligent control,
robotics and mechatronics including human-machine teaming, knowledge-based paradigms,
learning paradigms, machine ethics, intelligent data analysis, knowledge management, intelligent
agents, intelligent decision making and support, intelligent network security, trust management,
interactive entertainment, Web intelligence and multimedia.
The publications within “Advances in Intelligent Systems and Computing” are primarily
proceedings of important conferences, symposia and congresses. They cover significant recent
developments in the field, both of a foundational and applicable character. An important characteristic
feature of the series is the short publication time and world-wide distribution. This permits a rapid and
broad dissemination of research results.
Advisory Board
Chairman
Nikhil R. Pal, Indian Statistical Institute, Kolkata, India
e-mail: [email protected]
Members
Rafael Bello Perez, Universidad Central “Marta Abreu” de Las Villas, Santa Clara, Cuba
e-mail: [email protected]
Emilio S. Corchado, University of Salamanca, Salamanca, Spain
e-mail: [email protected]
Hani Hagras, University of Essex, Colchester, UK
e-mail: [email protected]
László T. Kóczy, Széchenyi István University, Győr, Hungary
e-mail: [email protected]
Vladik Kreinovich, University of Texas at El Paso, El Paso, USA
e-mail: [email protected]
Chin-Teng Lin, National Chiao Tung University, Hsinchu, Taiwan
e-mail: [email protected]
Jie Lu, University of Technology, Sydney, Australia
e-mail: [email protected]
Patricia Melin, Tijuana Institute of Technology, Tijuana, Mexico
e-mail: [email protected]
Nadia Nedjah, State University of Rio de Janeiro, Rio de Janeiro, Brazil
e-mail: [email protected]
Ngoc Thanh Nguyen, Wroclaw University of Technology, Wroclaw, Poland
e-mail: [email protected]
Jun Wang, The Chinese University of Hong Kong, Shatin, Hong Kong
e-mail: [email protected]
Dharmendra Saini
Editors
Cyber Security
Proceedings of CSI 2015
123
Editors
M. U. Bokhari Dharmendra Saini
Department of Computer Science Bharati Vidyapeeth’s College
Aligarh Muslim University of Engineering (BVCOE)
Aligarh, Uttar Pradesh New Delhi
India India
Namrata Agrawal
National Institute of Financial Management
Faridabad, Haryana
India
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
part of Springer Nature
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Preface
The last decade has witnessed remarkable changes in the IT industry, virtually in all
domains. The 50th Annual Convention, CSI-2015, on the theme “Digital Life” was
organized as a part of CSI@50, by CSI at Delhi, the national capital of the country,
during December 2–5, 2015. Its concept was formed with an objective to keep ICT
community abreast of emerging paradigms in the areas of computing technologies
and more importantly looking at its impact on the society.
Information and Communication Technology (ICT) comprises of three main
components: infrastructure, services, and product. These components include the
Internet, infrastructure-based/infrastructure-less wireless networks, mobile termi-
nals, and other communication mediums. ICT is gaining popularity due to rapid
growth in communication capabilities for real-time-based applications. New user
requirements and services entail mechanisms for enabling systems to intelligently
process speech- and language-based input from human users. CSI-2015 attracted
over 1500 papers from researchers and practitioners from academia, industry, and
government agencies, from all over the world, thereby making the job of the
Programme Committee extremely difficult. After a series of tough review exercises
by a team of over 700 experts, 565 papers were accepted for presentation in
CSI-2015 during the 3 days of the convention under ten parallel tracks. The
Programme Committee, in consultation with Springer, the world’s largest publisher
of scientific documents, decided to publish the proceedings of the presented papers,
after the convention, in ten topical volumes, under ASIC series of the Springer, as
detailed hereunder:
1. Volume 1: ICT Based Innovations
2. Volume 2: Next Generation Networks
3. Volume 3: Nature Inspired Computing
4. Volume 4: Speech and Language Processing for Human-Machine
Communications
v
vi Preface
We also take the opportunity to thank the entire team from Springer, who have
worked tirelessly and made the publication of the volume a reality. Last but not
least, we thank the team from Bharati Vidyapeeth’s Institute of Computer
Applications and Management (BVICAM), New Delhi, for their untiring support,
without which the compilation of this huge volume would not have been possible.
Chief Patron
Patrons
Prof. S. V. Raghavan
Department of Computer Science, IIT Madras, Chennai
Prof. Ashutosh Sharma
Secretary, Department of Science and Technology, Ministry of Science of Technology,
Government of India
Prof. K. K. Aggarwal
Founder Vice Chancellor, GGSIP University, New Delhi
Prof. M. N. Hoda
Director, Bharati Vidyapeeth’s Institute of Computer Applications and Management
(BVICAM), New Delhi
ix
x The Organization of CSI-2015
Advisory Committee
Editorial Board
A. K. Nayak, CSI
A. K. Saini, GGSIPU, New Delhi
R. K. Vyas, University of Delhi, New Delhi
Shiv Kumar, CSI
Anukiran Jain, BVICAM, New Delhi
Parul Arora, BVICAM, New Delhi
Vishal Jain, BVICAM, New Delhi
Ritika Wason, BVICAM, New Delhi
Anupam Baliyan, BVICAM, New Delhi
The Organization of CSI-2015 xi
xiii
xiv Contents
Contributors
xvii
xviii Editors and Contributors
C. Aka Assoua Anne-Marie AIIT, Amity University, Noida, Uttar Pradesh, India
Basit Ansari Marathwada Institute of Technology, Aurangabad, India
Anupam Baliyan Bharati Vidyapeeth’s Institute of Computer Applications
(BVICAM), New Delhi, India
Monika Bansal Rukmini Devi Institute of Advanced Studies, Delhi, India
Tosal Bhalodia Atmiya Institute of Technology and Science, Rajkot, India
M. U. Bokhari Department of Computer Science, Aligarh Muslim University,
Aligarh, India
Varun Chauhan Knowledge Graph Department, Binary Semantics Pvt. Ltd.,
Gurgaon, India
Krishna Keerthi Chennam Gitam University, Computer Science Engineering,
Hyderabad, Telangana, India
Kiran Chhabra Computer Science and Engineering, Dr. C.V. Raman University,
Bilaspur, CG, India
Prachi Dewal C-DAC, Noida, India
Nilanjan Dey Department of Information Technology, Techno India College of
Technology, Kolkata, India
Sangeeta Dhall Faridabad, Haryana, India
Bhawna Dhruv Amity University Noida, Noida, India
G. Dileep Kumar Bharathiar University, Coimbatore, India
Amit Doegar Department of CS NITTTR, Chandigarh, India
Avijit Dutta NIC, New Delhi, India
Ekta Department of CSE and IT, Bhagat Phool Singh Mahila Vishwavidyalaya,
Sonipat, India
Princy George Department of Computer Science and Engineering, SCMS School
of Engineering and Technology, Ernakulam, Kerala, India
Sakshi Goel Amity School of Engineering and Technology, Amity University,
Noida, India
B. B. Gupta Department of Computer Engineering, National Institute of
Technology Kurukshetra, Kurukshetra, Haryana, India
Himanshu Gupta AIIT, Amity University, Noida, Uttar Pradesh, India
Hina Gupta Amity School of Engineering and Technology, Amity University,
Noida, India
Editors and Contributors xix
Abstract Mobile gadgets today are swaggering computing potential and memory
at par or at times even higher to that found in desktop personal computers.
A wireless interconnection has turned out to be considerably more readily acces-
sible these days. As individuals are growing mobile with regard to the fast lifestyle
and working pattern, a new, smarter system came into existence that is termed as
“location-based service” (LBS). Such a system amalgamates the location data of a
user with smart applications to deliver demanded services. Although LBSs provide
major openings for a large variety of markets and remarkable convenience to the
end user, it also presents subtle privacy attack to user’s location information. Threat
to the privacy sneaks into the system due to the prerequisite of sending user’s
current location to the LBS provider to attain related services. Since the volume of
data gathered from dynamic or stationary mobile users using LBS can be high, it is
vital to outline the frameworks and systems in a manner that is secure and keep the
location information private. This can be portrayed as a big mobile data challenge in
LBSs setting. This paper aims to explore the issues related to privacy involved in
LBSs. In the paper, we introduce framework structure outline for preventing
location-based vicinity inference of users who issue a query and also proposed VIC-
PRO algorithm which helps to overcome the gaps of well-established K-anonymity
approach in the existing system. The suggested approach strengthens the privacy of
query initiating vicinity information.
1 Introduction
2 Motivation
With the continual reduction in the price of mobile devices, it is noticed that not
only the use of the location-aware gadgets raises in a growing number of civilian
and military applications, additionally a developing interest for regularly being
informed while out on the road for innumerable purposes. Keeping track of the
traffic condition, route information, on the fly parking information, en route grocery
store information, meeting a friend on way back home, and catching new movie in
theaters are few of such applications. Considering the metropolitan zone with
hundreds and thousands of vehicles (especially in a profoundly populated continent
like Asia) where every driver or passenger is interested in such information relevant
to their trips to plan visits more smartly and save their time in wasteful driving.
Such era of voluminous data can be viewed as big mobile data challenge in
LBSs-enabled mobiles.
Another major motivation behind writing the paper on this subject is the news of
November, 2014, where New York City Mayor declared that an association of four
companies named City Bridge will develop and manage up to 10,000 IEEE 802.11
access points for New York City’s LinkNYC [2]. It agrees to be the biggest free
municipal Wi-Fi operation in the world. In the same motion, the Prime Minister of
India announced to develop intelligent cities having geo-spatial mapping, Wi-Fi
hotspots, and intelligent transit system with GPS features. In both the mentioned
declarations, sharing user’s location information would play a major role in order to
access the demanded services. Clearly, LBS will be having a sweeping impact of
the digital world in the future as pointed out by the market analysis [3] and would
reach $63 billion by 2019.
3 Related Work
A survey of literature in the related field has brought forth several architectures,
algorithms and techniques that have been proposed by many authors in which they
have discussed about anonymity based, different cloaking mechanisms based and
trusted third party based privacy preservation models. A location estimation
enabled smart mobile device allows users to submit location-based queries to
web-based LBSs. Once the mobile apparatus throws the service request, the sender
has no control over the facts contained by the submitted query. An observer with a
right to access the information included in the query may utilize that information to
guess the user’s location. This makes a profound challenge of location privacy
protection that must be ponder upon. In this concern, most of the previous work
4 R. Gupta and U. P. Rao
4 Problem Formulation
5 Proposed Approach
In our proposed framework, after accepting the location information from the
sender, anonymizer runs proposed algorithm (VIC-PRO) and instead of including
K − 1 more users of same vicinity, this algorithm computes K users after per-
forming the following geometric transformation techniques and produces a final
anonymized query set Q:
a. Translation
b. Reflection
Figure 3 shows the instance after computation of suggested geometric
transformations.
The algorithm computes diverse K − 1 values assuming the nearest beacon node
as the center of origin. Each new direction is now having the same probability
considered to be the query initiator vicinity by an adversary. Anonymizer forwards
this anonymized query to LBS provider and after processing, the result set is
communicated back to anonymizer. Now, anonymizer has the actual result and
some false hits. Anonymizer filters out the incorrect results and sends the genuine
result to the end client.
6 Vic-Pro Algorithm
Initially K=0
1. Anonymized_query_set Q = Empty Set
Let, current_loc = (x, y)
2. x' = Reflection(x) and y’ = Reflection(y)
//Reflection method computes reflection geometric
//transformation for the given input point
3. Anonymized_query_set Q = Q U { x’, y’}
4. while (K<=18) // as K= 20 is assumed
5. {Select random translation factors δtx and δty
6. new_x = x + δtx
7. new_y = y + δty
8. Anonymized_query_set Q = Q U {new_x, new_y}
9. Increment K by 1}
10. Anonymized_query_set Q = Q U {current_loc (x,
y)}
11. return Anonymized_query_set Q
Privacy Protection Through Hiding Location Coordinates … 7
7 Example
The example shows the research gap in K-anonymity concept. SVNIT is taken as
the query originating region. Consider the geographical context as depicted in
Fig. 5.
Considering the case where a SVNIT student is generating a query asking for a
“Nearby 34 in. by 48 in. poster printing shop”. In K-anonymity principle, K users
become the part of anonymized query. Anonymizer includes K-1 more client from
the same vicinity and after that advances the anonymized inquiry to a service
(a)
(b)
(c)
(d)
provider. Here, in this situation, though the location and user information are
preserved, but breach of vicinity information privacy can be registered by an
adversary.
Following TWO cases, adversary may use to estimate region information:
Case1: Adversary staying informed with client’s social networking status If an
adversary knows that SVNIT is the only technical institute in the vicinity and he is
also keeping track of client’s social networking whereabouts in which client
mentioned that she is going to attend “Security & Privacy symposium-2015” and
later would also introduce a publication on a certain date at certain place. For an
adversary, with the help of other promotional information shared by the tagged
institute, which is organizing the mentioned event on other social connecting sites,
it would not be troublesome to discover that the student-initiated nearest poster
printing shop search query indeed belong to SVNIT.
Case2: Adversary staying educated with demographics Considering adversary
is now aware about the fact that Surat is a place known more for businesses and
individuals do not go for advanced education much. In such a case, it is not difficult
for an adversary with legitimate access to query information to figure out that the
query initiating region is SVNIT. If an adversary further probe with the help of
social networking connection, then user identification is also possible. This iden-
tification breach makes this framework susceptible to background knowledge
attack.
Figure 6a–d depicts few diverse regions produced after the suggested geometric
transformations at anonymizer.
8 Conclusions
This paper proposes a strategy and an algorithm VIC-PRO that computes the
anonymized query with K users present in diverse directions unlike the existing
approaches where the focus was on the inclusion of K clients from the same
vicinity. Suggested approach strengthens the framework and deals with preserving
vicinity privacy along with user identification privacy as each new direction now
has the same probability considered to be a query initiator vicinity.
References
4. Mokbel MF, Chow CY, Aref WG (2006) The new Casper: query processing for location
services without compromising privacy. In: Proceedings of the 32nd international conference
on very large data bases: VLDB endowment, pp 763–774
5. Mokbel MF (2007) Privacy in location-based services: state-of-the-art and research directions.
In: Proceedings of international conference on mobile data management IEEE, pp 228–228
6. Ghinita G, Kalnis P, Skiadopoulos S (2007) MOBIHIDE: a mobile peer-to-peer system for
anonymous location-based queries. In: Proceedings of 10th international symposium on
advances in spatial and temporal databases, Springer, Berlin, Heidelberg, pp 221–238
7. Gruteser M, Grunwald D (2003) Anonymous usage of location-based services through spatial
and temporal cloaking. In: Proceedings of the 1st international conference on mobile systems,
applications and services ACM, pp 31–42
8. Kalnis P, Ghinita G, Mouratidis K, Papadias D (2007) Preventing location-based identity
inference in anonymous spatial queries. IEEE Trans Knowl Data Eng 19(12):1719–1733
9. Chow CY, Mokbel MF, Liu X (2006) A Peer-to-Peer Spatial Cloaking Algorithm for
Anonymous Location-based Service. In: Proceedings of the 14th annual ACM international
symposium on advances in geographic information systems ACM, pp 171–178
10. Bettini C, Mascetti S, Wang XS, Freni D, Jajodia S (2009) Anonymity and
historical-anonymity in location-based services, in privacy in location-based applications.
Springer, Berlin Heidelberg, pp 1–30
11. Buchanan WJ, Kwecka Z, Ekonomou E (2013) A privacy preserving method using privacy
enhancing techniques for location based services. Mobile Netw Appl 18(5):728–737
12. Damiani ML, Bertino E, Silvestri C (2010) The probe framework for the personalized
cloaking of private locations. Trans Data Priv 3(2):123–148
13. Hein GW, Kneissl F, Avila-Rodriguez J-A, Wallner S (2005) Authenticating GNSS—global
navigation satellite systems—proofs against spoofs
Advanced RSA Cryptographic
Algorithm for Improving Data Security
Mukesh Kumar
Abstract Data security is a method which is used to cover the important infor-
mation. Data security methods control the privacy and integrity of the important
information. The access to the database of the companies has improved. Now
companies store their business data more on computer than before. Most of the
company data is for internal use and not for the general public because business
data is highly confidential. At present, cryptographic block cipher is being used
with some logical operation and the main drawback in this method is the generation
of the secret key which is totally based on the alphabets. So with the help of loop
concept, there is a chance for the hackers to find out the secret key. But I propose
advanced algorithm for cryptography which is totally dependent on hashing func-
tion technique to generate a secret key which is further used to encrypt and decrypt
the important information. The secret key will be generated by using different key
generation algorithms which will be of higher sets of alphanumeric characters. I am
using a hashing technique for cryptography along with a new quantum-bit gener-
ation method.
Keywords Cryptography Ciphertext Data security Key generation algorithms
1 Introduction
M. Kumar (&)
H.P. University, Shimla, India
e-mail: [email protected]
their data secret. At present, hackers are using a variety of techniques in order to
break into computer system and steal information or change important data of any
organizations (Fig. 1).
At present, cryptography has many applications area. Cryptography provides a
high level of privacy by covering confidential data of any individuals or groups.
Cryptography is used for the purpose of providing access to data in restricted way,
data integrity, and authentication. At present, a lot of research work is going on to
find out the new cryptographic algorithms based on security and complexity [1].
Simply talking about data security, then the following features come to our mind
like privateness, validation, wholeness, disownment, access control, and availability
of data.
2 Literature Review
For generating the starting key for the purpose of encryption and decryption of the
information/ message provided to use, they generally used the random key gen-
erator algorithm [2]. In that particular technique, a replacement technique is used
where they can take four words set from given input message and after getting the
encrypted data, the equivalent words in the random key matrix can be identified.
A technique is suggested by Nath in multiple sequence alignment algorithm for
searching characters from a random key matrix [3]. In this technique, they provide
arrangement for encrypting data. The random key matrix contains all possible
characters set whose value lies between 0 and 255 (ASCII code). Text/ word key
used by the end user is used for the design of the random key matrix. From the
Advanced RSA Cryptographic Algorithm for Improving Data Security 13
starting word/text key used by user, they are providing a new encryption/
decryption steps to find the random and encryption number. The author unex-
pectedly finds it very difficult to meet the two variables from two unlike input. At
this stage, it is very difficult to crack the encryption techniques which are provided
by the author, but if anybody wants to crack the techniques then they must know all
the possible pattern of the word/text key used by the user. If anybody tries to
decrypt the data, then they must know the exact key matrix used and if anybody
tries theoretically to make the random matrix then they almost have to try 65,536!
attempts. Different researcher have applied this technique on possible data files and
they have observed that it gives 100% results while applying cryptography on data.
In paper [4], a newly advanced algorithm outline symmetric algorithm is pro-
posed which is in resemblance to Rijndael algorithm. In Rijndael technique, 128
bits block for encryption are used but in AES technique, 200 bits block are used.
3 Rivest–Shamir–Adleman Algorithm
Quantum-bits Production: For the secret data of the user, first of all, find the
secret key for that data to encrypt. After getting the secret key, change the data into
hexadecimal code and after that change it into binary to get the least bits as
Quantum bits of 0 and 1.
4 Proposed Work
5 Conclusion
This presented algorithm is mainly used for block cipher techniques and possibly
this technique will take less time to encrypt a data of large size. It is not possible to
crack the encryption algorithm if you do not know the exact key value and this is
the main advantage of presented algorithm. I have used this algorithm for both
encryption and decryption while sending or receiving important message.
References
1. A Text book by William Stallings, data and computer communications, 6e William 6e 2005
2. Chatterjee D, Nath J, Dasgupta S, Nath A (2005) A new symmetric key cryptography
algorithm using extended MSA method: DJSA symmetric key algorithm, accepted for
publication in IEEE CSNT2011 to be held at SMVDU (Jammu) 03–06 June 2011
3. Nath A, Das S, Chakrabarti A (2010) Data hiding and retrieval. In: Proceedings of IEEE
international conference on computer intelligence and computer network held at Bhopal from
26–28 Nov 2010
4. Muhammad F, Chowdhury I, Matin MA Effect of security increment to symmetric data
encryption through AES methodology, 9th ACIS international conference on software
engineering, artificial intelligence, networking, and parallel/distributed computing. Phuket,
Thailand. 6–8 Aug 2008. https://doi.org/10.1109/SNPD.2008.101
5. Felten K An algorithm for symmetric cryptography with a wide range of scalability. In
published by 2nd international workshop on embedded systems, internet programming and
industial IT
6. Wang Y, Hu M (2009) Timing evaluation of the known cryptographic algorithms. International
conference on computational intelligence and security. Beijing, China, 11–14 Dec 2009.
https://doi.org/10.1109/CIS.2009.81
7. Nath A, Ghosh S, Malik MA Symmetric key cryptography using random key generator, vol 2,
pp 239–244
Author Biography
Abstract The main aim of this paper is to provide the security for accessing the
collaborative computing environment. Main thing here is using authentication
method, users can access their collaborative environments. So normal authentica-
tion is not sufficient for collaborative environment, that is why here I am proposing
two-factor authentication for collaborative environment.
Keywords Security Collaborative computing environment Authentication
Two-factor authentication
1 Introduction
1.1 Authentication
Something authentication is not enough for some type of environments, and then
we are using two-factor authentication mechanism. Main example for this one is
collaborative environment. First by using passwords, we can authenticate then the
environment cannot be connected here again we will authenticate by using the
advance techniques to access environments. Users having these two authentication
information only can access the data via environment [2].
• Face recognition
• Fingerprint recognition
• Smart cards
• Secret message
• One-time password (OTP)
Fig. 2 Recognition of
fingerprints
It is a charge card measured card with embedded consolidated circuits, which can
handle and recollect data. To validate, the client must swipe or physically embed
the card in a unique per user (Fig. 3).
Secret message authentication is the most useful technique. In this method, we are
using some secret message to access the data. Once the environment established,
some secret message created by using this message the credentials can be
authenticated.
Where stamping into PC systems and destinations is concerned, it for the most part
incorporates entering a superfluous mystery word, which is genuine just once and
changes every time you sign in. This is known as a one-time secret word and
another is delivered new every time you get to the system. Commonly a one-time
mystery word [7] is a movement of useful in vain numbers or characters or it might
be around six or something like that short, unpredictable words. How might you
know your one-time mystery word if it keeps developing? It is not something you
are obliged to remember that: it is delivered actually and sent to you by some
system other than online transmission. It might be sent to your wireless (cell phone)
as a SMS text [8]; it could be delivered by an application running on your phone or
by a conferred, handheld electronic contraption called a security token; it may even
be printed out and sent to you on paper, the colossal outdated way (Fig. 4).
22 G. Dileep Kumar and R. Praveen Sam
2 Conclusion
In this paper, just we have mentioned what all are the mechanisms available in the
collaborative environment accessing authentication methods. In future papers, we
have to discuss the particular two-factor authentication method and how the method
is implemented.
3 Future Work
References
1. Li Q, Cao G (2011) Multicast authentication in the smart grid with one-time signature. IEEE
Trans Smart Grid 2(4)
2. www.explainthatstuff.com-how-security-tokens-work.html
3. Nguyen MD, Bui QM Your face is NOT your password
4. Marcel S, Rodriguez Y Biometric face authentication using pixel-based weak classiers
5. Woodward JD Jr, Horn C, Gatune J, Thomas A A look at facial recognition
Different Security Mechanisms in Two-Factor … 23
6. FRVT 2006 and ICE 2006 Large-scale results. National Institute of Standards
7. Aloul F, Zahidi S, El-Hajj W Two factor authentication using mobile phones
8. Corella F, Lewison K Strong and convenient multi-factor authentication on mobile devices
Author Biographies
Abstract The growing connectivity across the globe has been made possible due
to Internet and web applications. Social networking and e-commerce web appli-
cations are dominating the cyberspace. Cybersecurity tends to secure the computer
system, information and the connecting network from the attackers whose intention
is to misuse the information and damage the resources. Network security is a bit
complex mechanism as compared to information security as it does not only require
securing end systems but also ensure the security of entire network system. This
paper discusses the network security measures and the change in trend in appli-
cation of these measures.
1 Introduction
With the latest innovations in the communication technologies, public networks are
now relied for sharing personal and financial information. Earlier, security protocols
were not developed for Internet to secure itself and were not even implemented in
TCP/IP. This left computer systems vulnerable to attacks. Cybersecurity consists of
measures for securing computer system, data and the associated network. The major
intention is to secure the network for communication.
Network security implements policies that prevent unauthorized access by
checking unique ID and password and monitors the requests for network resources
to prevent misuse of computer network and denial of genuine user requests.
The attributes that assure security in a network and the respective attack methods
that are used by the attackers to compromise the security of the network are listed in
Table 1.
1.2 Preparedness
In the recent past, many discussions have led to recognizing the potential threats
related to cyberspace and their impact on the information infrastructure and their
connecting network. These threats and the nature of attackers vary according to
problem domain and application area network and hence the consequences. The
impact varies from target system to the network ranging from nominal to severe.
Thus, preparedness against threats in cyberspace is mandatory and integral to
mission assurance [2].
2 Security Management
The precautions and planning to handle the attacks should be done at all levels
through long vision strategies which can be done by characterizing the threats that
are expected and the risks can be detected at the preliminary stages.
Cyberthreats for the network are oriented in five levels of preparedness as shown in
Fig. 1. The intent of damage increases with the level. The lower levels implement
the already existing security frameworks [3–6] and the higher levels follow mission
assurance strategy where the assets or the resources are preserved from attacks on
mission capabilities.
The security measures as a countermeasure for the techniques used by attackers for
respective cyberthreat levels are summarized in Table 2.
The cyberattacks are mainly focused for stealing or hacking the secure information.
The information that is available in cyberspace is present for knowledge sharing. It
is desired to be shared among authorized or allowed group of users and needs to be
protected from the unwanted and harmful adversaries. Saltzer–Schroeder [7] rec-
ognized the need for providing control on the access of information by the users.
Sandhu-White [8] discussed the limitations of already existing models for secure
information sharing, such as discretionary access control, mandatory access control
and role-based access control. They explained how these models are effective in
handling the issues what their main motive is but are not capable of responding in
scenarios where monitoring of life events related to cybersecurity is needed. They
proposed group-centric SIS models to address the limitations of traditional models
for such cases. Zhao-White [9] realized the importance of information sharing and
classified threat alert levels for cybersecurity. The proposal was to form a
‘Collaborative Information Sharing Framework’ that would enhance the pre-
paredness for cybersecurity. Zhao-White [10] also proposed a group-centric col-
laborative information sharing framework that aims to improve community
cybersecurity by analysing information sharing requirements in the community
through the designed formal policy model for the framework.
28 S. Maurya and A. Singhrova
3 Mission Centricity
A few years ago, the computer security was based on confidentiality, integrity and
availability (CIA) of IT assets. The network security combines authorization with
CIA triad to ensure security of the overall network. Network security elements like
honeypots and darknet analysers are used to control attacks trying to gain unau-
thorized access and misuse and manipulate network resources. They do so by
attracting harmful traffic away from the main computer resources and in the
meantime analysts get to track the attacker and as a result improve network security.
However, traditional systems have limited operability, and are not able to protect
the network in most of the cases as nowadays attackers have sufficient knowledge to
avoid getting trapped in the darknets and bypass the CIA triad for security.
Considering its loopholes and the failure in providing completely secure infras-
tructure, the need for ‘Mission Centric’ paradigm increased which should be able to
manage the rapid changes in the operational context and dynamic time and space
bound behaviour of missions [11].
‘Changing Trend in Network Security Measures: A Review’ 29
3.2 Adaptation
Network security measures focus on securing both private and public network.
Wireless network security, IP security, firewalls and physical security are the major
concerns of network security. Wireless security aims at preventing unauthorized
access or damage to computers through wireless networks. The most common types
of wireless security are Wi-Fi protected access (WPA) and wired equivalent privacy
(WEP) which are based on the concept of ‘Secure Connection’ and ‘Secure
Session’ [17]. Internet protocol security (IPsec) is a protocol for securing com-
munications by authenticating and encrypting each IP packet of a session. IPsec
includes protocols for establishing mutual authentication between agents at the
beginning of the session and negotiation of cryptographic keys to be used during
the session. A firewall acts like a barrier through which the traffic going in and out
of the network must pass. A firewall security allows only the traffic that is autho-
rized to pass. Packet filtering router, application-level gateway and circuit-level
gateway are the three types of firewalls implemented to secure the network [18].
Table 3 lists the advanced network security measure improvements that have
proven to prevent the network system from latest attacks.
5 Conclusion
The network security combines authorization with CIA triad to ensure security of
the overall network. But the traditional measures have failed in providing com-
pletely secure infrastructure from the modern attackers. Hence, the need for
‘Mission Centric’ paradigm increased which is capable in managing the rapid
changes in the operational context and dynamic time and space bound behaviour of
missions. The resilient network systems that are adaptive in nature are the need of
the time that should implement latest network security measures with advance
implementations and modifications.
References
1. Adeyinka O (2008) Internet attack methods and internet security technology, In: Second Asia
international conference on modeling & simulation, AICMS 08, pp 77–82, 13–15 May 2008
2. Bodeau DJ, Graubart R, Fabius-Greene J (2010) Improving cyber security and mission
assurance via cyber preparedness (cyber prep) levels, In: IEEE second international
conference on social computing (socialcom), pp 1147–1152, 20–22 Aug 2010
3. National Institute of Standards and Technology (NIST) (2009) Recommended Security
Controls for Federal Information Systems and Organizations, NIST Special Publication
800-53, Rev. 3, August 2009 (updated with errata May 1, 2010)
4. NIST (2010) Guide for applying the risk management framework to Federal information
systems: a security life cycle approach, NIST SP 800-37 Revision 1, February 2010
5. Information Technology (IT) Governance Institute (ITGI), COBIT (Control Objectives for IT
and Related Technology) (2007) V4.1
6. International Standards Organization (ISO) and International Electrotechnical Commission
(IEC), ISO/IEC 17799:2005 (2005) Information technology—security techniques—code of
practice for information security management
7. Saltzer JH, Schroeder MD (1975) The protection of information in computer systems.
Proc IEEE 63(9):1278–1308
8. Sandhu R, Krishnan R, White GB (2010) Towards secure information sharing models for
community cyber security. In: 6th international conference on collaborative computing:
networking, applications and worksharing (collaboratecom), pp 1–6, 9–12 Oct 2010
9. Zhao W, White G (2012) A collaborative information sharing framework for community
cyber security. In: IEEE conference on technologies for homeland security (HST), pp 457–
462, 13–15 Nov 2012
10. Zhao W, White G (2014) Designing a formal model facilitating collaborative information
sharing for community cyber security. In: 47th Hawaii international conference on system
sciences (HICSS), 2014, pp 1987–1996, 6–9 Jan 2014
32 S. Maurya and A. Singhrova
Author Biographies
Keywords Random early detection (RED) Active Queue Management (AQM)
IETF (Internet Engineering Task Force) ns-2 network simulator
K. Chhabra (&)
Computer Science and Engineering, Dr. C.V. Raman University,
Bilaspur, CG, India
e-mail: [email protected]
M. Kshirsagar
Yashwantrao Chawan College of Engineering, Nagpur, MS, India
e-mail: [email protected]
A. Zadgaonkar
Dr. C.V. Raman University, Kargi Raod, Bilaspur, CG, India
e-mail: [email protected]
1 Introduction
Due to the gigantic growth of the Internet, it faces the challenge of congestion [1,
2]. AQM is the solution to this congestion control. Its two main objectives are high
link utilization with low packet loss and low packet queuing delay. Recommended
by IETF [3] RED was first AQM, proposed by Sally Floyd and Jacobson in 1993
[4].
In RED algorithm, Exponential Weighted Moving Average (EWMA) filter as
given by Eq. (1) is used for averaging of current queue size. For detection of actual
congestion not transient one, averaging procedure is used. Here, two thresholds
THmin minimum threshold and THmax maximum threshold are used and the average
queue size is compared with these thresholds. If the average queue size is less than
THmin no packet is dropped and if it is greater than THmax every incoming packet is
dropped. When it lies between two thresholds, packets are randomly dropped using
drop probability calculated by Eqs. (2) and (3).
avgi þ 1 1wq avgi þ wq qi ð1Þ
where
avgi+1 average queue size at (i + 1)th time
avgi average queue size at ith time
wq moving weighted average constant
qi current queue size
THmin minimum threshold for queue
THmax maximum threshold for queue
Pmax maximum value for Pb
Pa current packet marking probability
Pb temporary probability used in calculation
count packets since last discarded packet
Being simple in its approach, RED achieves goal of congestion avoidance,
removes global synchronization, and overcomes drawback of Drop tail [3]. Despite
having been widely used in combination with TCP for several years, RED has not
found acceptance in the Internet research community [1]. It has some disadvantages
like hard parameter setting, congestion depends on parameters, insensitivity toward
input traffic and there is no matching between average queue length and current
queue length [1]. After RED’s proposal, lot of research is carried out to find its
suitability. Many variants of RED [2, 5–9] were proposed to overcome short-
comings of it with different congestion notification approaches used like the
average queue size, packet loss and link utilization and control theory, etc.
An Improved RED Algorithm with Input Sensitivity 37
The remainder of this paper is organized as follows. Section 2 gives the pro-
posed work along with approach used, Sect. 3 gives simulation results and Sect. 4
deals with conclusions and the final section gives all the references used.
2 Proposed Work
Due to the exponential growth of the Internet, AQM fails to respond fast to traffic
changes or to adapt to time-varying TCP dynamics. This motivated us for our
proposed work. We have tuned the threshold parameter of RED that is threshold
upper limit and lower limit of the average size in terms of router buffer size. Setting
THmin to an appropriate value could help router to make good link utilization.
Improper setting of THmax also causes more packets losses once the average queue
size exceeds it. We have tuned THmin to 40% of maximum queue size and THmax to
70% of it. This tuning has given us good results as well as it helps to achieve the
link utilization by properly keeping router buffer efficiently utilized. Since RED’s
performance degrades heavily shown by wild oscillations when the traffic load
becomes bursty indicating it is insensitive to input. We have incorporated this
problem in our approach. A little amount work is already done in [10]. LTRED
algorithm is shown in Fig. 1.
For each packet arrival, average value for the queue size is calculated using EWMA
as in original RED. To have impact of input traffic the current queue size is
compared with a threshold which is 75% value of maximum buffer value (qt1) and
depending upon that average value is updated by 5% of the current queue size. This
step has impact of input sensitivity. Then it proceeds as RED algorithm. In order to
find out if the congestion is sustained once again two checks are made. First, the
current queue size is compared with another threshold (85% value of maximum
buffer value—qt2) for bursty input and the average queue value is compared with a
reference value which is mid value of upper and lower threshold value. If this
comparison is found to be true, then packets are dropped earlier as compared to
RED algorithm and this is early congestion indication to sources so that they should
slow down their rate of sending packets. In rest part, it will continue to work as in
original RED.
38 K. Chhabra et al.
Where
avg : average queue size
qi : current queue size
B : maximum buffer size
q t1 : First threshold for q i
q t2 : Second threshold for q i
TH min : Minimum Threshold value for avg
TH max : Maximum Threshold value for avg
mid th : Reference Threshold value for avg
We have implemented the proposed scheme using network simulator ns-2 [11]. To
evaluate the improvements network topology is shown in Fig. 2. Bandwidth and
delay associated with links are also given in the diagram and Gs to Gr is bottleneck
link. We have compared proposed scheme LTRED with RED [4], ARED [7] and
AVQ [2], standard AQMs and plotted graphs for various cases.
Total simulation time period is 30 s for the topology shown Two FTP sessions
randomly start in between 0 and 0.01 s and lasts till the end. In the middle of
simulation, another “m” FTP session would randomly start in between 10.0 and
10.1 s which is to simulate changes of network conditions. TCP Reno is used for all
AQMs (RED, ARED, AVQ and LTRED). Parameters used are, for RED and
ARED are THmax = 15, THmin = 5, queue size q = 30, Maxp = 0.1, wq = 0.002,
and for ARED and AVQ other parameters are set as per default value as in ns-2.
In case of LTRED THmin = 12 and THmax = 21, qt1 = 22.5, qt2 = 25.5, and
midth = 16.5 rest settings are similar to RED. We have observed results for sources
varying “m” from 30 to 60 (Tables given) and drawn different graphs for 60 sources
An Improved RED Algorithm with Input Sensitivity 39
Senders Receivers
S1 R1
2Mb & 10ms 2Mb & 10ms
S2 R2
0.7Mb
Gs R3
S3 Gr
10Mb 20ms 10Mb
• •
• •
Sn
14 to 15ms 14 to 15ms Rn
for all the approaches used. We have observed performance in case of packet loss,
packet arrival ratio, the average queue size, and the current queue size.
Figure 3 shows the graph for comparison of packet loss for all the four cases
showing minor differences during initial period of less traffic. After 10 s when
traffic increases that is during congestion period LTRED shows less number of
packet losses overall, and shows improvement by 2.3–4.3%. Figure 4 shows total
number of packets received for all cases, it shows that packets delivered number is
largest in case of LTRED, showing improvement in throughput by 2.19–3.48%. In
Fig. 5, the graph shows effective congestion notification to sources in terms of
packet arrival rate. Sources are earlier informed about congestion and they reduce
their sending rate here also LTRED outperforms other AQMs.
Figure 6 depicts current queue changes, showing more variation in transient
congestion and later on in highly congestion area, AVQ uses less buffer size Wild
oscillations in RED and ARED show unstable behavior and in case of LTRED
nearly stable behavior. Figure 7 illustrates bandwidth utilization which shows that
LTRED outperforms other AQMs. Figure 8 shows average queue variation for
RED, ARED, and LTRED, EWMA approach is used in these cases to calculate the
average queue size. In this case also LTRED outperforms RED and ARED by being
within its reference limit and depicts stable behavior. Figures 9, 10, and 11 show
comparison of the average and the current queue size for RED, ARED, and LTRED
cases. Out of these in LTRED case, mismatch behavior of the average and the
current queue size is reduced a lot. Tables 1, 2, 3, and 4 give simulation results for
number of sources varying from 30 to 60, showing superiority of LTRED.
An Improved RED Algorithm with Input Sensitivity 41
4 Conclusion
In this work, we have proposed an AQM, LTRED which is sensitive to input traffic
and also tuned threshold parameter in terms of buffer size, comparing it with RED,
ARED and AVQ have demonstrated better performance in terms of less number of
packet loss, effective congestion indication, high throughput, reduction in mismatch
behavior of the average size and the current size and high goodput values due to
which retransmission of packets also reduces a lot, and effective buffer utilization at
44 K. Chhabra et al.
router. An adaption mechanism based on input traffic is designed to drop the packet.
The key concept is that as traffic load changes and queue length deviates from a
threshold, dropping of packet occur.
References
1. Ryu S, Rump C, Qiao TC (2003) Advances in internet congestion control. IEEE Commun
Surv Tutorials 5(1):28–39. https://doi.org/10.1109/COMST.2003.5342228
2. Kunniyur S, Srikant R (2004) An adaptive virtual queue [AVQ] algorithm for active queue
management. IEEE/ACM Trans Networking 12(2):286–299
3. Braden B, Clark D et al. (1998) Recommendations on queue management and congestion
avoidance in the Internet. IETF Request for Comments RFC 2309
4. Floyd S, Jacobson V (1993) Random early detection gateway for Congestion avoidance.
IEEE/ACM Trans Netw 1(4):397–413
5. Sun J, Ko K, Chen G, Zukermam M (2003) PD-RED: to improve the performance of RED.
IEEE Commun Lett 7(8):406–408
6. Athuraliya S, Li V et al. (2001) REM: active queue management. IEEE Netw 15(3):48–53.
https://doi.org/10.1109/65.923940
7. Floyd S, Gummadi R, Shenkar S (2001) Adaptive RED: an algorithm for increasing the
robustness of RED’s active queue management. Berkely CA [online]. http://www.icir.org/
floyd/red.html
8. Li M, Zhao W (2010) Representation of a stochastic traffic bound. IEEE Trans Parallel Distrib
Syst 21(9):1368–1372
9. Wang H, Ye Z, Wang B (2011) Using auto-tuning proportional integral probability to
improve random early detection. IEEE 13th international conference on communication
technology (ICCT)
10. Chhabra K, Kshirsagar M, Zadgaonkar A (2015) Effect of load and threshold variation on
performance of RED: random early detection. Int J Sci Res 4(6):2319–7064. ISSN (online)
11. NS [network simulator] (1999) [online]. Available http://www.isi.edu/nsnam/ns
Author Biographies
Kiran Chhabra received her B.E. in Computer Science from Nagpur University in 1995 and M.E.
in Computer Technology and Applications from Chhattisgarh Swami Vivekanand Technical
University, Bhilai in 2009. She is pursuing her Ph.D. in Computer Science and Engineering in
congestion avoidance area. Currently, she is working as Associate Professor in Computer Science
Department at MM College of Technology, Raipur.
Dr. Manali Kshirsagar received her B.E. in Computer Technology from Nagpur University in
1992, M.E. in Computer Science and Engineering from Amravati University in 2001. She was
awarded Ph.D. in 2009 by Allahabad University for her work on the Data Mining Strategy to
explore Cotton Genome. Currently, she is working as Professor and Head of the Computer
Technology Department at Yashwantrao Chavan College of Engineering, Nagpur. She has many
An Improved RED Algorithm with Input Sensitivity 45
papers to her credit in various international journals, the international conference, and national
conference. Her areas of interest include Data Mining, Biometrics, and Computer Networks. She is
also a member of professional bodies like MIE, ISTE, and ACM.
Dr. A. S. Zadgaonkar has obtained B.E. in Electrical Engineering from Pt. Ravishankar Shukla
University, studying at Govt. Engineering College, Raipur in 1965. He obtained M.E. in 1978
from Nagpur University. His research paper for M.E. was awarded “Best paper” by the Institution
of Engineers [India] in the year 1976 and 1977 respectively. The testing technique for the quality
of wood developed by him was included in ISI in 1979. He was awarded Ph.D. in 1985 by Indira
Gandhi Kala & Sangeet University, Khairagah for his work on “Acoustical and Mechanical
Properties of Wood for Contemporary Indian Musical Instrument Making.” He was awarded
Ph.D. in 1986 by Pt. Ravishankar Shukla University on “Investigation of Dynamic Properties of
Non-Conducting Materials Using Electrical Analogy.” He has 47 years of teaching experience. He
has published more than 500 technical papers in various journals, and National and International
conferences. He has written four books on Engineering and five on “Science and Technology in
Indian Mythology.” He is currently adding glory to the post of Vice Chancellor of Dr. C. V.
Raman University, Bilaspur[Chhattisgarh]. He is life member of Acoustical Society of India,
Biomedical Society of India, Linguistic Society of India, Indian Society for Technical Education
and many social bodies
Security Attacks in Wireless Sensor
Networks: A Survey
Abstract Security is one of the major concerns in sensor networks. Wireless sensor
network comprises of huge amount of nodes called as tiny sensor nodes. The nodes
are required to exchange information with different nodes via wireless links in short
intervals. The information may be potentially private regarding people and business
processes. These networks suffer from adversary due to distributed behavior and
deployment in distant areas. The networks are governed by some constraints at
sensor node level like less battery power, less memory capacity, and low trans-
mission range while at network level, they are governed by ad hoc networking and
irregular connectivity. The paper analyzes the challenges, main security issues,
security breaches in wireless sensor networks and lists their defensive measures.
1 Introduction
Wireless sensor networks (WSN) are self-configured network with tiny sensor
nodes. Each wireless node possesses low energy, memory space, and computational
power. Components of sensor node include front end of radio, microcontroller,
main power supply, and sensors. The task of sensors is to monitor physical and
environmental conditions such as humidity, pressure, sound, temperature, and many
more. After monitoring, they send data to their main location. The data is requested
on basis of these parameters in sensor networks (Fig. 1).
Wireless sensor network has resource constraints that act as hindrance in using
existing security approaches. In fact, threats in sensor networks in context of
routing are susceptible due to simple routing protocols. There are many obstacles in
security for example, limited resources, unreliable communication, and unattended
operation.
This paper is organized as follows: In Sect. 2, literature survey is presented.
Section 3 discusses various protocols for wireless sensor networks. Section 4 gives
the security framework discussing standard goals, constraints, obstacles, security
breaches, and security mechanism. Finally, concluding the paper in Sect. 5 with
conclusion and security threats in different protocol layer along with their defensive
measure.
2 Literature Review
The nature of wireless sensor networks includes multiple nodes that make system
vulnerable to adverse effects and loss of information. It has led to look into security
and privacy aspects of networking by introducing new insistent technologies like
wireless sensor networks [1]. Author proposed a security framework for wireless
sensor networks, i.e., adaptive security architecture. It includes low-, medium-, and
high-level security modes. SENP protocol is used which aims at securing patterns
providing authenticity, confidentiality, and integrity. Agent-based secure routing
scheme involves use of trusted neighbors that is proposed in [2] which employs use
of probability and MAC model for identifying trustworthy neighbors, through
which secure routes are set up. The work given in [3] presents a group-based
security scheme for wireless sensor networks which includes sequential procedure:
Cryptographic key pre-distribution, group-based deployment, secure data aggre-
gation and rekeying. The author in [4] employs use of virtual grid connection to
secure data from end to end at multiple base stations. In revocation scheme, random
polynomial is used. The author in [5] proposed security protocol for verifying
model, i.e., TinySec + LEAP. TinySec holds binary operations, viz., authentication
and semantic secured encryption. In [6], author proposed improved fiestal based
ciphers for WSN. All of WSN’s block ciphers are designed using a 16 round fiestal
data block. Author proposes to use controlled permutation boxes for implementa-
tion of a fiestal scheme. The author in [7] proposed two secure and efficient data
transmission(SET) protocols for cluster-based wireless sensor networks, called
SET-IBS and SET-IBOOS, y using the identity-based digital signature(IBS)
scheme and the identity-based online/offline digital signature scheme, respectively.
In [8], the author presented an architecture utilizing concept of autonomic com-
puting and a simple object access protocol (SOAP) based interface to metadata
access points (IF-MAP) external communication layer to create a network security
sensor. A flexible two-level communication layer based on autonomic computing
and service oriented architecture is presented. In [9], the author proposed an
adaptive specification based intrusion detection system (IDS) for detecting mali-
cious unmanned air vehicles (UAVs). An IDS audits UAVs in a distributed system
to determine if the UAVs are functioning normally or are operating under malicious
attacks. In [10], the author proposed a realistic and reliable IDS architecture for the
advanced metering infrastructure (AMI). An AMI system is responsible for col-
lecting, measuring, and analyzing energy usage data and transmitting this infor-
mation from a smart meter to a data concentrator and then to a headend system in
the utility side.
3 WSN Protocols
Protocols are the set of rules and communication standards that must be followed by
source and destination in order to communicate with each other. There are several
types of communication protocols which can be grouped into the lower level,
high-level, and application-based protocols. Example includes TCP/IP that is set of
protocols consisting of more than 65,000 protocols (Tables 1 and 2).
Table 2 Classification of TCP/IP IP, TCP, UDP, SMTP, POP3, RIP, FIP, DHCP
protocols [11]
Cellular GPRS, GSM, WAP AND CDMA
VOIP SPX, RIP, MEGACO, MGCP AND H.323
General Frame relay, ATM, X.25, PPP
4 Security in WSN
4.2 Constraints
4.3 Obstacles
See Fig. 8.
Security Attacks in Wireless Sensor Networks: A Survey 55
5 Conclusion
There is a need for effective security mechanisms in wireless sensor networks. The
paper describes constraints, goals, obstacles, and security breaches based on dif-
ferent protocol layers, defensive measures, and security mechanism for wireless
sensor networks. The attacks in protocol layer and their measures are shown in
Table 3.
Table 3 Security threats in different protocol layers along with their defensive measures
Protocol Security breaches Defensive measures
layer
Physical Jamming Spread spectrum (FHSS)
layer Tampering Eavesdrop on the wire which is between memory chip
and microcontroller
Data link Collision Spread spectrum
layer Exhaustion Limit the mac admission control rate
Network Selective Multipath routing
layer forwarding Encryption
Acknowledgement Cryptographic methods
spoofing Four-way handshaking scheme
Black hole Identity verification protocol
Wormhole Validation technique
Hello flood
Sybil
Transport Flooding Bidirectional verification
layer Desynchronization Authentication
Application Data aggregation Encryption
layer Distortion Confidentiality protection
Clock skewing Synchronization protocols
56 P. Dewal et al.
References
1. Prasad NR, Alam M (2006) Security framework for wireless sensor networks, Springer
2. Devanagavi GD, Nalini N, Biradar RC (2014) Trusted neighbour based secured routing
scheme in wireless sensor networks using agents. Springer, New York
3. Hamid MA, Sarkar AMJ (2011) A group based security scheme in wireless sensor networks.
Springer
4. Ferng H-W, Nurhakim J, Horng S-J (2013) Key management protocol with end to end data
security and key revocation for a multi-BS wireless sensor network. Springer, New York
5. Tobarra L, Cazorla D, Cuartero F, Diaz G, Cambronero E, Model checking Wireless sensor
network security protocols: TinySec + LEAP*, Spain
6. Pazynyuk T, Li J-Z, Oreku GS (2008) Improvrd Feistal based ciphers for wireless sensor
network security. J Zhejiang Univ 9(8):1111–1117
7. Huang L, Li J, Guizani M (2014) Secure and efficient data transmission for cluster-based
wireless sensor networks. IEEE Trans Parallel Distrib Syst 25(3):750–761
8. Vollmer T, Manic M, Linda O (2014) Autonomic intelligent cyber-sensor to support
industrial control network awareness. IEEE Trans Ind Inf 10(2):1647–1658
9. Mitchell R, Chen I-R (2014) Adaptive intrusion detection of malicious unmanned air vehicles
using behavior rule specifications. IEEE Trans Syst Man Cybern Syst 44(5):593–606
10. Faisal MA, Aung Z, Williams JR, Sanchez A (2015) Data-stream-based intrusion detection
system for advanced metering infrastructure in smart grid: a feasibility study. IEEE Syst J 9
(1):31–44
11. Stavron E, Wireless sensor network, part 2: limitations. http://webhosting.devshed.com/c/a/
Web-Hosting-Articles/Wireless-Sensor-Networks-part-2-Limitations/
12. Fatema N, Brad R (2013) Attacks and counterattacks on wireless sensor networks. Int J
Ad-Hoc Sens Ubiquit Comput 4(6):1–15
13. Hu Y-C, Perrig A, Johnson DB (2006) Wormhole attacks in wireless senor networks. IEEE J
Sel Areas Commun 24(2):370–380
14. Padmavathi G, Shanmugapriya D (2009) A survey of attacks, security mechanisms and
challenges in Wireless sensor networks. Int J Comput Sci Inf Secur 4(1 & 2):1–9
15. Xiong NN, Cheng H, Hussain S, Qu Y (2013) Fault tolerant and ubiquotous computing in
sensor networks. Int J Distrib Sens Netw 2013:2. Article ID 524547
16. Christin D, Rosskopf C, Hollick M, Martucci L, Kanhere S (2012) IncogniSense: an
anonymity-preserving reputation framework for participatory sensing applications. In:
Proceedings of the IEEE international conference on pervasive computing and communica-
tions, pp. 135–143
17. Akyildiz I, Su W, Sankarasubramaniam Y, Cayirci E (2002) A survey on sensor networks.
IEEE Commun Mag 40:102–114
18. Cristofaro E, Soriente C (2013) Participatory privacy: enabling privacy in participatory
sensing. IEEE Netw 27:32–36
19. Erfani S, Karunasekera S, Leckie C, Parampalli U (2013) Privacy-preserving data aggregation
in participatory sensing networks. In: Proceedings of the 8th IEEE international conference on
intelligent sensors, sensor networks and information processing, pp. 165–170
20. Sharifnejad M, Shari M, Ghiasabadi M, Beheshti S (2007) A survey on wireless sensor
networks security. SETIT
21. Cardenas AA, Berthier R, Bobba RB, Huh JH, Jetcheva JG, Grochocki D, Sanders WH
(2014) A framework for evaluating intrusion detection architectures in advanced metering
infrastructures. IEEE Trans Smart Grid 5(2):906–915
22. Vollmer T, Manic M (2014) Cyber-physical system security with deceptive virtual hosts for
industrial control networks. IEEE Trans Industr Inf 10(2):1337–1347
Security Attacks in Wireless Sensor Networks: A Survey 57
Author Biographies
Vishal Jain has completed his M.Tech. (CSE) from USIT, Guru
Gobind Singh Indraprastha University, Delhi and doing Ph.D. in
Computer Science and Engineering Department, Lingaya’s
University, Faridabad. Presently, he is working as Assistant
Professor in Bharati Vidyapeeth’s Institute of Computer
Applications and Management, (BVICAM), New Delhi. His
research area includes Web technology, Semantic Web, and
information retrieval. He is also associated with CSI, ISTE.
58 P. Dewal et al.
1 Introduction
Secure transaction of data in real time is always needed for various confidential
business operations or many other private data sharing operations. For keeping these
transactions and communications safe from the intruders, cryptography is one of the
most approached techniques. There are two types of techniques that are used for
encryption and decryption—the first one is a symmetric key cryptography and the
second one is asymmetric key cryptography. Encryption may be achieved by two
types of ciphering schemes—stream cipher and block cipher as mentioned in [1].
First of all, the concept cellular automata (CA) was proposed by von Neumann
[2]. In the past two decades, many areas of applications of cellular automata
(CA) have been explored by many researchers as mentioned in [3–6]. Applications
of cellular automata are used in different fields like—physics, chemistry, mathe-
matics, biology, computer science, communication and engineering, etc. Recently,
it has been applied in BioHash code generation [7], image encryption [8] and
ECG-Hash code generation [9], watermarking [10], and authentication.
Cellular automata (CA) have some specific characteristics like—balancedness,
correlation immunity, nonlinearity, easy to implement, etc. These characteristics
satisfy the essential cryptographic properties. In this paper, a new idea is repre-
sented in which a CA-based cryptosystem is generated. This cryptosystem shows
the high quality of randomness of the patterns which have similar significances
regarding the older computational techniques of cryptography [11]. And further
enhancement of quality of randomness can be embedded with the help of using
programmable cellular automata (PCA) [11, 12]. The suggested cryptographic
technique in this paper uses a single block of 1-D PCA.
The draft of the paper is organized as follows. In the current section, discussion
is on how CA and PCA can be corresponded with important cryptographic features.
Section 2 contains some basics of cellular automata (CA) and important termi-
nologies of CA. Section 3 presents the proposed encryption and decryption algo-
rithms using PCA theory. Section 4 shows experimental results. Section 5
concludes our work and suggests the future scope of the work.
Cellular automata are dynamical system in which space and time are discrete that
operate according to local interaction rules [2].
Here in our experiment, an example of a 1-D cellular automata (CA) is con-
sidered. It has two possible states per cell, i.e., S = (0, 1) and 3 neighborhoods [3].
Each cell in a CA is updated based on its old state and the state of its left and right
neighbors. The basic model of a CA is shown above in Fig. 1. In general, state of a
CA at any time instant t is represented as a vector as
Symmetric Key Encryption Technique … 61
st ¼ ðx1 ; x2 ; . . .; xn Þ; ð1Þ
where xi denotes the bit in the ith cell xi at time instant t and n denotes the length of
the bit string. The bit in ith cell at the “next” time instant t + 1 is given by fi. It is
known as transition function. For elementary CA transition function for next state
of each cell is denoted as
xti þ 1 ¼ f i xti1 ; xti ; xti þ 1 where i ¼ 2; 3; . . .; n 1
NeighborhoodState
Fig. 1 Model of CA
A binary cellular automata (CA) containing n cells as x1, x2, …, xn is null boundary
when the left neighbor of leftmost cell and right neighbor of rightmost cell is taken
as “0”. The CA is called periodic boundary CA when the neighbor of the extreme
cells is considered as adjacent to each other.
Concepts of PCA were introduced into [1] and its hardware implementation was
realized into [11]. A PCA is an altered cellular automata (CA) that contains control
signals and applied combinational logic in it is indefinite. Each cell in PCA is
controlled by these control signals. With the help of these control signals, different
rules can be applied on a single 1-D structure dynamically. Figure 2 shows the
model of a PCA, which consists of three control signals. By using these signals, any
possible combination of complemented and non-complemented additive CA rules
[1] can be implemented. A PCA has high degree of flexibility, parallelism, and
complexity, which redefines cellular automata (CA) as more appropriate and robust
for cryptographic applications.
In this paper, a symmetric key cryptographic technique using PCA has been
implemented. In this experiment, non-complemented rule 150 and rule 102 are used
to form a group cellular automata [1] of cycle length 12. These cycles of even
length are used for encryption and decryption. Block of plain text is loaded into
PCA and after six cycles, the ciphertext is generated and after next six cycles, the
ciphertext is decrypted to retrieve plain text. Permutation group representations of
Symmetric Key Encryption Technique … 63
CA having even cycles of even length are mentioned in [1, 13]. The procedure for
cryptography, which has been followed in the proposed work is formulated as
Input: Rule 90, 102, and 150 are applied on each cell of PCA and selection of
these rules depends on the selection lines of multiplexer (Fig. 3).
Step 1: When selection lines S0S1 are logic “00” then rule 90 is selected.
Step 2: When selection lines S0S1 are logic “01” then rule 150 is selected.
Step 3: When selection lines S0S1 are logic “10” or “11” then rule 102 is
selected.
64 D. Parashar et al.
Fig. 3 A 4 1 multiplexer
for scheduling rules
In this paper, 1-D cellular automata were explored in order to recognize the use of
cellular automata in the field of cryptography. A PCA-based symmetric key
cryptographic technique is proposed which uses the block cipher scheme. It can
provide a good level of security and safety of data over Internet. In future, more
cellular automata rule combinations can be explored in order to identify more
applications in various fields of technology. Furthermore, in future this work can be
used to implement the image encryption technique using MATLAB. It can also be
used in 2-D cellular automata and parallel programming, which can reduce the
Symmetric Key Encryption Technique … 67
complexity than that of the proposed scheme. This work can be used in
steganography as well. The hardware implementation of this technique is easy and
economical. This makes the technique also applicable in embedded system designs
and other computer science technologies.
References
1. Nandi S, Kar BK, Chaudhuri Pabitra Pal (1994) Theory and applications of cellular automata
in cryptography. IEEE Trans Comput 43(12):1346–1356
2. von Neumann J (1966) In: Burks AW (ed) Theory of self reproducing automata. Univ. of
Illinois Press, London
3. Kotoulas L, Tsarouchis D, Sirakoulis GC, Andreadis I (2006) 1-d cellular automata for
pseudo random number generation and its reconfigurable hardware implementation. In:
Proceedings of IEEE international symposium on circuits and systems
4. Choudhury PP, Sahoo S, Chakraborty M, Bhandari S, Pal A (2009) Investigation of the global
dynamics of cellular automata using boolean derivatives. Int J Comput Math Appl 57:1337–
1351
5. Wolfram S (2002) A new kind of science. Wolfram Media Inc. ISBN: 1-57955-008-8
6. Wolfram S (1986) Theory and applications of cellular automata. Wolfram Scientific
7. Dey N, Nandi B, Dey M, Das A, Chaudhuri SS (2013) BioHash code generation from
electrocardiogram features. In: 3rd IEEE international advance computing conference
8. Nandi S, Roy S, Dey N, Nath S, Chakraborty S, Kaara WBA (2014) 1-D group cellular
automata based image encryption technique. In: IEEE international conference on control,
instrumentation, communication and computational technologies (ICCICCT), pp 578–583
9. Nandi S, Roy S, Dansana J, Kaara WBA, Ray R, Chowdhury SR, Chakraborty S, Dey N
(2014) Int J Comput Netw Inf Secur
10. Acharjee S, Chakraborty S, Ray R, Nath S, Dey N (2014) Watermarking in motion vector for
security enhancement of medical videos. In: International conference on control, instrumen-
tation, communication and computational technologies (2014)
11. Anghelescu P (2012) Hardware implementation of programmable cellular automata
encryption algorithm. In: IEEE international conference on telecommunication and signal
processing, Prague, pp 18–21
12. Anghelescu P, Sofron E, Rîncu C, Iana V (2008) Programmable cellular automata based
encryption algorithm. Semicond Conf 2:351–354
13. Wolfram S (1985) Cryptography with cellular automata. In: Proceedings of the conference,
CRTPTO’85 on advances in cryptography. Lecture notes in computer science, vol 218,
pp 429–432
14. Ross SM (2010) Introductory statistics, 3rd edn. Academic Press, Elsevier, pp 676–681
A Comparative Study on Lightweight
Cryptography
1 Introduction
As the current scenario, use of smart devices such as credit card, smart card,
personal digital assistant (PDA), RFID tags, wireless sensor nodes, etc., is gaining
equipotent role in our daily life. Their use is much ubiquitous. On the other hand,
security and performance is one of the severe issues for such devices. So, we can
need considerable security as well as performance of these devices, owning mini-
mal storage space, and computational capabilities. This results in raising a research
area known as lightweight cryptography. The aim of LWC is to provide the secured
information on highly constrain relevant devices owning minimal sources. LWC
algorithms have got certain common features like they must possess low power
consumption, lesser communication cost, low area, low energy, as well as little
processing time. The implementation of LWC is done in such a way that it increases
throughput and efficiency. The ubiquitous use of RFID tags rise concern about
equipotent security in RFID system. Since low-cost tags are extremely
resource-constrained device, common security approach is no longer applicable to
them. Hence, one challenging topic is to purpose a secure lightweight cipher that is
suited for RFID tags. This paper describes a comparative study among some
well-known lightweight ciphers (Table 1).
2 Stream Cipher
3.1 BSF-128
The BSF-128 stream cipher is designed on the basis of grain. It has designed for
128-bit secret key applications. BSF-128 consists two shift registers, one FCSR and
one LFSR of 128-bit length each. It also uses an S-Box of 8 16, i.e., it takes 8-bit
input and produced output of 16-bit. The S-Box is a combination of Skipjack and an
S-Box designed by ISRC at QUT, which has also been used in SOBER t-16 cipher.
On the basis of cryptanalysis, we assume that this cipher is secure against many
cryptanalysis attacks [2].
3.2 Grain
RC4 designed by Ron Rivest in 1987. In the history of cryptography, RC4 has been
one of the most popular stream ciphers. Its internal state contains a permutation of
72 M. U. Bokhari and S. Hassan
bits overall possible bytes sequence ranging from 0 to 255. Its design analysis and
approach are quite different as compared to LFSR-based stream ciphers. The
internal state consists a table of N ¼ 2n , n-bit words and two n-bit pointer [7, 8].
There are some attacks on RC4 based on the relationships between the internal
states of the S-Boxes [9].
WG-7 [10] stream cipher is based on the primitive WG stream cipher [11]. It is
designed by Y. Luo, “Q. Chai”, “G. Gong”, and “X. Lai” in 2010. WG-7 is a
very fast stream cipher for the lightweight devices (for example smart mobile
phone, RFID tags, as well as wireless sensor node) WG-7 is design. Both WG
and WG-7 are hardware-oriented stream cipher that uses a word-oriented LFSR
and a filter function based on WG. WG works on GFð229 Þ but WG-7 in GFð27 Þ.
WG-7 uses 80-bit secret key and 81-bit IV and the LFSR is clocked 46 times, the
internal state consists of 161 bits and the security level claimed by the designer is
80-bits [4, 12].
Recently, a distinguishing attack was discovered against the WG-7 stream cipher.
Within the time complexity of Oð227 Þ, an attacker can recover both the secret key
and internal state of the cipher [13].
3.6 HITAG2
A cryptanalysis against HITAG2 was founded [15], which easily broke HITAG2 by
a SAT solver within several hours. Besides the brute force attack, this is only a
unique cryptanalysis on HITAG2 that break the security of cipher. This attack
comprises of three phases [16].
• To extract 32 bit of secret key a black box attacks is vulnerable.
• To achieve other key bits the white-box attack seems also vulnerable.
• Brute force searches for the remaining key bits.
• Cost-optimized parallel code-breaker COPACOBANA is able to reveal the
secret key of a HITAG2 transponder in less than 2 h (103.5 min) in the worst
case [17].
3.7 SOBER
SOBER is a popular family of stream ciphers that are widely used in embedded
devices. It was first proposed by G. Rose in 1998. Their family includes several
stream ciphers: Sober t-16, sober t-32 [18], sober t-128 [19], and many more. The
synchronous stream cipher sober t-16 and sober t-32 were submitted to NESSIE
program [12] with 128-bit key and 256-bit key strength, respectively. Almost
all ciphers, which belong, to sober are depending on similar principle and virtu-
ally have equivalent model structure. Most of the sober family ciphers consist of
three basic components [19].
• Linear feedback shift registers (LFSR)
• Nonlinear function
• Stutter control.
74 M. U. Bokhari and S. Hassan
The physical unclonable function (PUF) [20] is promising solution to mitigate the
effect of physical attacks. PUF is the physical entity that generates output based on
their input and intrinsic physical properties of embedding hardware. It exploits only
those physical properties of embedding devices.
A block cipher works on two pair of algorithms, one for encryption e and other for
decryption d. A group of plain text namely P of size L [ 1 are encrypted together
by the encryption function C ¼ eðk; PÞ; C yield cipher text under the enciphering
function e with key k. A whole block of size L is encrypted with a single key k at a
time. The key k is a composition of several values ki ki1 . . .k1 k0 . After encryption,
the ciphertext C is decrypted by the set of composite key k under the correspon-
dence P ¼ dðk; CÞ. In this cipher, the ciphertext block is totally depends upon the
key k. In mathematical terminology, we can also say that
eðk; PÞ ¼ d 1 ðk; PÞ ¼ C
dðk; PÞ ¼ e1 ðk; CÞ ¼ P
5 DESL
The algorithms tend to make use of inside this cipher are the DESL (build up
extension of DES) & ECC, DESL is the enhanced lightweight version of DES. The
microchip size of DESL is considerably reduced as individual S-Box is utilized
frequently for eight times. Which can make DESL compact, tough, efficient, and
prevented from linear and also differential cryptanalysis attacks. The DESL is
proficient to enciphering 64 bit of plain text in 144 clock rounds while it is working
with a frequency of 100 kHz and gaining current of 0.89 lA [22]. The DESL
structure involves mainly the building blocks controller, mem-left, mem-right, key
program, and S-Box.
In 1985, two American mathematicians Vector Miller and Neal Koblitz proposed
the concept of elliptic curve cryptography. Their theory is completely based on
elliptic curve discrete logarithms and NP hard problem that requires a complete
exponent time. The application of ECC involves information security, personal
digital assistant (PDA), wireless communication network, wireless sensor nodes,
image encryption, smart cards, e-commerce, and also in economic-based commu-
nication protocols. ECC makes use of 162-bit public key by the assist of pick-
ing points on elliptic curves, and afford a security strength that is corresponding
to 1024-bit key in RSA [23].
• For higher security (Largest ECC & RSA system broken to date are 108-bit
512-bit)
• Largest effort ever expanded in PKC challenge for solving 108-bit ECC.
Amount of work required was about 50 times of 512-bit RSA.
8 Conclusion
WG-8 LW stream – 80 80 Able to resist AA, COA, DFA, No High throughput 20 stage LFSR XOR and
cipher CUA, DGA, DFTA, TMDA multiplication
WG LW stream – 128 128 Able to resist AA, COA, DFA, No – 32 stage LFSR XOR and
16 cipher CUA, DGA, DFTA, TMDA multiplication
RC4 Stream No 128 – Some attack based on relationship Yes Best No XOR
cipher between internal state and S-Boxes
Sober Stream – 256 – Secure, have a good immune Yes – LFSR XOR
t-32 cipher
HC Hybrid 16 128– – Most secure, and immune to most 44 Maximum LFSR XOR
cipher 256 of LA, DA, CUA, BDA, AA etc.
AA Algebraic attack; BDA Birthday attack; BFA Brute force attack; COA Correlation attack; CUA Cube attack; DA Determine attack; DFA Differential attack;
DFTA Discrete Fourier transformation attack; DGA Distinguish attack; DMA Davies Murphy attack; GDA Guess &determine attack; HB Hummingbird; IV
Initialization vector; KRA Key recovery attack; LA Linear attack; LC Linear cryptanalysis; LW Lightweight; SCA Side channel attack; TA Timing attack;
TMDA Time–memory–data attack
77
78 M. U. Bokhari and S. Hassan
References
S. N. Panda
Abstract Wireless sensor networks are associated with assorted functional aspects
including battery or energy, power, log of neighboring nodes, cache, and number of
services. In a network attack, the malicious node or packet attempts to temporarily
or permanently halt these parameters so that the authentic and realistic communi-
cation can be damaged. Such attacks were previously associated with DDoS attacks
which do not allow the authentic user to access the services. Number of algorithms
devised against DDoS attacks but very less treatment to the vampire attacks which
is more hazardous as it is very difficult for the authentic user to confirm whether
there is any attack on network. It consumes battery of node very rapidly which is
not identified by the network node. In our proposed algorithm, a unique and
effective algorithm for location-based key generation is devised and implemented
which makes use of dynamic key exchange based on the location.
Keywords Vampire attacks Wireless sensor network security
Reliable communication Energy optimization in wireless sensor network
1 Introduction—Vampire Attacks
S. N. Panda (&)
Chitkara University, Rajpura, Punjab, India
e-mail: [email protected]
2 Features
Vampire attack is not definite to any specific protocol or topology or port. Such
attacks do not interrupt the services directly. It affects the resources using the
services. The vampires make use of protocol compatible or compliant messages
vampire attacks transmit data that drain the energy level of nodes. Vampires do not
change or interrupt altering discovered paths or routes.
Resource Draining Attacks—Such attacks create and send the assaults or attacks
which means the creation and sending the messages by spiteful node by which the
energy consumption is taken a lot by the assault.
Carousal Attack—These attack drastically increases the routing path and length
which creates delay in the networks and also inadequate by the number of allowable
entries in the resource route [2] (Figs. 2 and 3).
Stretch Attack [3]—Such attacks creates the artificial or fake routes using which
the packet can be disguised. The attack forces the data packet to choose the fake
path so that there is huge delay and battery consumption increases. Stateless
GPS Hash Table Based Location Identifier Algorithm … 83
protocols are source routing protocols that keep track of the communication in the
network infrastructure [4]. Here, the source node mentions the complete and whole
route to the destination that is inside the packet header. Intermediaries do not create
the independent or arbitrary forwarding decisions. In stateful protocols, the nodes
have advance information of state, topology, forwarding techniques and routes.
Network nodes create the local forwarding decisions on that stored state.
These attacks are basic in remote wireless sensor systems. They are normally
alluded as way-based Denial of Service (DoS) attacks [6]. Restricted hash chains
can keep these attacks by constraining the rate at which hubs transmit packets.
In wormhole attack, the vindictive hub builds a passage (way) to the destination in
such a path, to the point that all the bundles from the source are exchanged by
means of the assailant which can change substance of the bundle before sending it
to the destination. To keep this attack, parcel rope system is utilized [8]. A chain is
data added to the bundle in order to confine the parcel’s voyaging separation.
Fig. 4 Proposed model for avoidance of vampire attacks in wireless sensor networks
120
20
0
1 2 3 4
Simulation Attempt
Fig. 5 Comparison between and classical and proposed in terms of energy consumption
88 S. N. Panda
6 Conclusion
In the proposed work, to avoid and detract the vampire attacks, an effective
location-based identifier is integrated in the network that will generate a dynamic
key based on the GPS location and current timestamp. Using this approach, the
genuine packets shall not be lost. The packet loss is associated with the malicious
node. In the simulation scenario, the overall integrity and reliability of the network
are improved using the proposed algorithmic approach.
References
1. Vasserman EY, Hopper N (2013) Vampire attacks: draining life from wireless ad hoc sensor
networks. IEEE Trans Mob Comput 12(2):318–332
2. Vijayanand G, Muralidharan R (2014) Overcome vampire attacks problem in wireless ad-hoc
sensor network by using distance vector protocols. Int J Comput Sci Mob Appl 2(1):115–120
3. Manimala S, Devapriya AT (2014) Detection of vampre attack using EWMA in wireless ad
hoc sensor networks. IJISET Int J Innovative Sci Eng Technol 1(3):450–550
4. Khanna MMR, Divya S, Rengarajan A (2007) Securing data packets from vampire attacks in
wireless ad-hoc sensor network. Int J Innov Res Comput Commun Eng 2 (An ISO 3297: 2007
Certified Organization)
5. Kaul S, Samuel H, Anand J (2014) Defending against vampire attacks in wireless sensor
networks. Int J Commun Eng Appl IJCEA 5, Artical C084, March
6. Anand J, Sivachanda K (2014) Vampire attack detection in wireless sensor network. Int J Eng
Sci Innov Technol (IJESIT), 3(4), July
7. Channawar PM, Chavan YV (2015) Vampire attack: energy efficient trust based solution. 3
(7), July
8. Chandekar MRS, Nayyar V (2014) Defending against energy draining attack in ad-hoc
sensing network. 1(V1), November
9. Chumble MSC, Ghonge MM, Mitigation of vampire attack in wireless ad-hoc sensor network
10. Guptha NS, Lavanya NL, Detection and mitigation of vampire attacks in wireless ad-hoc
sensor networks
11. Raikar MR (2014) Prevention of vampire attacks to control routing behavior in wireless ad
hoc sensor networks
12. Du R, Chen C, Yang B, Lu N, Guan X, Shen X (2015) Effective urban traffic monitoring by
vehicular sensor networks. IEEE Trans Vehicul Technol 64(1):273–286
13. Karlof C, Wagner D (2003) Secure routing in wireless sensor networks: attacks and
countermeasures. Ad hoc networks 1(2–3):293–315
GPS Hash Table Based Location Identifier Algorithm … 89
14. Du W, Deng J, Han YS, Chen S, Varshney PK (2004). A key management scheme for
wireless sensor networks using deployment knowledge. In: INFOCOM 2004. Twenty-third
Annual Joint Conference of the IEEE computer and communications societies (vol 1). IEEE,
March
15. Conti M, Di Pietro R, Mancini L, Mei A (2011) Distributed detection of clone attacks in
wireless sensor networks. IEEE T Depen Sec Comput 8(5):685–698
16. Liu A, Ning P (2008) TinyECC: A configurable library for elliptic curve cryptography in
wireless sensor networks. In: Proceedings of the 7th international conference on Information
processing in sensor networks (pp 245–256). IEEE Computer Society, April
17. Othman SB, Bahattab AA, Trad A, Youssef H (2014) Secure data transmission protocol for
medical wireless sensor networks. In: Advanced Information Networking and Applications
(AINA), 2014 IEEE 28th International Conference on (pp 649–656). IEEE, May
18. Jayasudha AC, Venkatesh V (2014) Energy competent cluster based prediction framework for
wireless sensor network
19. Zhao J, Yagan O, Gligor V (2014) On topological properties of wireless sensor networks
under the q-composite key pre-distribution scheme with on/off channels. In: Information
Theory (ISIT), 2014 IEEE International Symposium on (pp 1131–1135). IEEE, June
Abstract Cloud computing (CC) has emerged as the next generation of computing
in IT Enterprise. Earlier, IT services are under proper physical and personnel
control, CC migrates the application softwares and databases to the huge data
centers, where the management of data along with services is done by the cloud
provider. However, this characteristic poses many security challenges which have
not been well understood. This paper focuses on security of data at cloud storage,
which has always been a most important issue in CC security. Many approaches
have been proposed to protect data in cloud which are not sufficient to meet the
requirements of a cloud user. So in this paper, a model for data security at cloud
storage site has been proposed.
1 Introduction
Cloud computing (CC) is a general term refers to outsourcing of hosted services and
computing resources over the Internet. Here, resources refer to network resources,
virtualized servers, platforms, computing infrastructures, etc. Business profession-
als routinely face several business-related problems. CC adopts concepts from
service-oriented architecture (SOA) that can help the business professionals to
outsource resources and application software’s in the form of services. CC provides
its resources as services using well-established standards and best practices to allow
on-demand and broad network access to cloud services.
Various Internet-based online services like Amazon EC2, Amazon S3, etc., do
provide tremendous amounts of storage space and computing resources. Storing
data into the cloud storage offers great advantage to its tenants since they do not
have to worry about the complexities involved in management of hardware or
software. Since all the data maintenance tasks are done by cloud storage provider,
this eliminates the burden of local machines. Although CC is a promising service
platform for the Internet technologies, this new computing environment brings
about many challenging issues which had profound influence on its adoption. One
of the biggest issues is cloud data storage security. As the amount of data is
growing, as are the growing need for security. Biggest concerns with cloud data
storage is that of data integrity, confidentiality, and availability at untrusted servers.
The main aim of this paper is to extend/improve the existing three-dimensional
algorithms [1, 2]. Prasad et al. [1] had not included integrity constraint in their
proposed formula for classification and data stored on cloud storage is not in
encrypted form. Authors titled their work as three-dimensional but only two
dimensions (confidentiality and availability) were involved in their proposed work.
In [2], there was no mechanism for addressing the issue of data integrity and user
authorization is done by the cloud provider, due to which there was data owner’s
loss of control issue. This paper is an attempt to overcome these limitations.
The rest of the paper is organized as follows. Section 2 summarizes related
work, Sect. 3 provides the detailed motivation behind this work and Sect. 4 dis-
cusses the data security model for cloud, Sect. 5 provides a comparative analysis of
proposed technique with the existing techniques and Sect. 6 gives some concluding
remarks.
2 Related Work
Kulkarni et al. [3] had proposed a CC framework which works in two phases
namely data classification and 3D accessibility. During data classification phase,
user’s data is classified based on CIA (Confidentiality, Integrity, and Availability)
parameters specified by the user during storage. After classification and using their
proposed formula, the priority rating of data is computed. The data having higher
rating is considered as critical and hence 3D security is recommended for that data.
In accessibility phase, authors had used OTP and two-factor password techniques to
avoid data leakage and impersonation. The author does not provide basis for pro-
posed formula and value of integrity constraint is not used anywhere in the pro-
posed formula.
Prasad et al. [1] and Deokar et al. [4] had proposed the similar technique for
classification of data as proposed in [3]. After classification of data in [1], the user
who wants to retrieve data needs to register first and for every data access user’s
identity is authenticated for certain authorizations whereas in [4], different password
scheme for distinct category of data were proposed. Same problems were identified
in the proposed work as in [3] along with the limitation that the data stored at cloud
Data Security Model in Cloud Computing Environment 93
storage site is unencrypted form and if the username and password are compro-
mised, the data can easily be retrieved by any malicious entity.
Wang et al. [5] had proposed a data integrity verification scheme using homo-
morphic token and erasure codes which facilitate the integrity checking of data
stored at distributed sites. The proposed methodology also supports dynamic data
operations of updation, insertion, and deletion. Their experimental analysis illus-
trates that the proposed scheme is secure against Byzantine failure and can identify
inadvertent data modifications.
Wang et al. [6] had proposed to introduce a third-party auditor (TPA) to check
the integrity of stored data on behalf of cloud user. The client leverages its few
privileges to TPA to audit the correctness of its data due to lack of computational
power. In prior works, most data integrity checking schemes fail to verify data
correctness after performing data modification operations (update, insert, and
delete). Their proposed scheme allows facility to check integrity of updated data
also.
Lijo et al. [7] had presented a user-centric solution for data security. According
to this scheme, a client agent is incorporated to manage the activities of data
auditing on behalf of Cloud user. One of the main limitations of this scheme was
that cloud provider’s approval is required to incorporate client agent in cloud
application. If there is lot of inter-cloud communications involved, then this solu-
tion could not be applied because it further requires consent of other cloud pro-
viders also. As data stored is in encrypted form, the client must be equipped with
some computational power to do encryption and decryption tasks.
In [2], Tirodkar et al. had presented a scheme for data categorization as proposed
in [1], and afterwards, for different category of data a distinct user authentication
scheme was used.
3 Motivation
In the work done by Prasad et al. [1], the following limitations were observed. They
had calculated criticality rating of data on the basis of values passed by user for
respective CIA parameters. The proposed formula was s½i ¼ ðC½i þ ðA½i 10Þ=2,
which do not incorporate integrity parameter. They had also not provided any basis
to derive formula. Further data stored at cloud storage site is in unencrypted form
and if the username and password are lost, the data could easily be accessed by any
malicious user. The formula is supposed to derive CIA parameters in order to be
called as three-dimensional but formula uses two parameters of confidentiality and
availability only. In the work done by Tirodkar et al. [2], they implemented three
different user authentication mechanisms for different categories of data but there is
no provision for addressing the issue of data integrity. Once data is stored at cloud
storage, data owner lost its control over it and cloud storage provider can manage
data leakage to help rival parties.
94 M. Kumari and R. Nath
To address the limitations mentioned in Sect. 3, this paper presents a model for data
security for effective data security in CC. The paper presents improved mechanism
of data handling. The improved technique works in three phases—categorization,
storage, and retrieval. In the following sections, abovementioned phases were
discussed in detail.
As all data sent to the cloud for storage is not equally sensitive, hence a uniform
level of security is not advantageous. Therefore, the data should be categorized
based upon the level of its sensitivity.
When a client wants to upload data for storage in the cloud, he has to provide the
values of confidentiality (C), integrity (I), and availability (A) parameters in the
scale of 1–10. The value of C parameter is based on extent of secrecy required,
value of I parameter is based degree of assurance of accuracy is needed and value of
A is based on how often data is accessed.
The sensitivity rating of the data is computed using the following equation:
Depending upon the value of SR, data are classified into three categories—
public, confidential, and sensitive. If SR value lies between 1 and 3 (1 and 3
inclusive) then data is labeled as public. If SR value lies between 4 and 6 (4 and 6
inclusive) then data is labeled as confidential and if SR value is greater than 6 then
data is labeled as sensitive.
Algorithm for Categorization of Data
Data_Label[x] = “Confidential”
ELSE
Data_Label[x] = “Sensitive”
End If
End For
3. Output: D [ ], Data_Label [].
After successful storage of data by cloud provider, there must be secure procedures
for retrieval as well. When a user wants to retrieve data, it has to make a request to
cloud storage provider and needs to register with the owner/organization. After
successful registration user gets a username and a password. This generated user-
name and password are necessary for retrieval of data and at the same time this
username is forwarded to cloud storage to store into its directory for future
transactions.
Table 1 shows the authentication mechanism for a client to have access to a
different category of data [2] depending upon the sensitivity level as characterized
by categorization phase.
The following paragraphs discuss the mechanism of retrieval of data belonging
to different categories, viz., public, confidential, and sensitive.
a. Retrieval of Public Data
When a user wants to retrieve data which belongs to public data category then
user has to register itself if he is already not registered.
User has to send a request along with its registered username to have access to
data. The cloud provider first checks in, to which category requested data
belongs. The cloud provider first checks the username into its stored directory, if
the username does not match with any entry it asks the user for registration. If
the username matches with one entry in directory then it redirects the request to
data owner for authentication as depicted in Fig. 3. After successful authenti-
cation, user will be allowed to access public data.
decryption keys and associated digital signature. User transmits digital signature
to cloud storage provider. These digital signatures also act as authentication
means to authenticate the user to cloud storage provider. Cloud storage provider
verifies the digital signature and allows access to data. By using the issued
decryption key user can access data.
The following guidelines were followed regarding data access:
• A user granted access on public data is not allowed to access confidential and
sensitive data.
• A user granted access on sensitive data is allowed to access confidential and
public data.
The latest two approaches reported in the literature are [1] and [2]. The proposed
approach is compared with the existing two approaches proposed by Prasad et al.
[1] and Tirodkar et al. [2]. The comparison has been made on the following
parameters—identification and authentication, confidentiality, integrity, availabil-
ity, non-repudiation, encryption, and security if user identity and password are
compromised and security from cloud provider.
The comparison shows that the proposed model fulfills all the parameters listed
in Table 2, while Prasad et al. [1] approach fulfills only four parameters and
Tirodkar et al. [2] fulfills only five parameters.
6 Conclusion
This paper has presented a data security model based on the data sensitivity level.
The model has classified the entire data to be stored in the cloud into three cate-
gories—public, confidential, and sensitive. Depending upon the category, different
kinds of security mechanisms are applied. The proposed model has been compared
with the existing two approaches and has been found much better than other
approaches. In addition, as the user authentication is done by the data owner itself,
it mitigates the issue of loss of control to a much extent and data owner can keep
track of who has accessed its data.
References
Author Biographies
1 Introduction
The security remains the biggest issue in Cloud computing, as Cloud provides
services which are located on a remote location and it is the trust of consumer on
Cloud service providers that their data will be secured. The integrity and confi-
dentiality of the user’s data are at risk as they do not have physical control over the
data [3]. This is because of the fact that the Cloud server is at different location and
client is located at different locations. Therefore, the server cannot be trusted com-
pletely for managing details of users and access rights. The data of user is at risk due
to insider attacks or compromised servers. This can be overcome if the users trust
Cloud service providers to secure and properly manage their data [4]. Due to the fact
that the insiders are very well familiar with the infrastructure, procedures of oper-
ation and terms and conditions of the organization the attack of malicious insiders
are more severe [4]. Insider attacks are done by malicious employees at any location,
i.e., provider’s or user’s. The attacks caused by the insiders have an adverse effect on
the trust of the Cloud user on the provider. Passwords, cryptographic keys, and files
can easily obtained by the malicious insider. These attackers not only damage the
financial value but also the reputation of an organization [4].
HIDS cannot be used as attackers may not leave traces in the operating system of
the host where the IDS are residing. NIDS cannot detect the attack if the com-
munication is encrypted. In clouds, distinct users share various resources. The
attacks can migrate from and be intended for any of the Cloud resources. Thus only
DIDs can be used. But, the challenges in the adoption of the DIDS in Clouds are
(i) Distinct types of users and user requirements; (ii) Complex architecture; and
(iii) Different requirement of security. MA-based IDS are not suitable as
(i) Hierarchical structure poses problem of reliability and scalability, and (ii) Not
flexible to protect from the attacks on IDS itself. GIDS are not suitable as (i) Every
service model (SaaS, IaaS and PaaS) has different set of threats, users, and
requirements; (ii) Clouds are highly scalable; (iii) GIDS solution cannot correlate
the alerts from the different nodes; (iv) Performance and load balancing are needed
more in Clouds than Grid [6].
4 Literature Review
Rule-based learning for the identification of insiders and a solution for the detection
of wrong insider activities have been given by the authors [7]. Some of the threats
identified include insecure shared technology vulnerabilities, application pro-
gramming interfaces, and malicious insiders. When an attack occurs, machine
learning techniques are used to raise an alarm. The seven common activities of the
insiders are (Table 1):
For activity classification purposes, the following machine learning techniques
are used (Table 2).
It has been found by the analysis that decision tree C4.5 and multilayer per-
ception are better for activity classification in Cloud-based environment. The result
of the confusion matrix reveals C4.5 as the best classifier.
Authors in [8] have used technique of multithreading for improving the per-
formance of the IDS. NIDS proposed sensitizes as well as monitors the network
traffic using the sensors. In this model, the Cloud user accesses the remote servers
over the Cloud network. The monitoring and logging of the requests and the actions
of the user are done by a multi-threaded NIDS, which has large data handling
capacity and also reduces the packet loss.
104 P. Oberoi and S. Mittal
A combined approach for malware detection and root kit prevention used in [9]
in virtualized Cloud environment. The IDS is intended to execute on VM instances
with a backend Cloud to share out malware scanning operations among numerous
back ends. Flexible, distributed security solution is given with a minimal overall
resource footprint on the end host. The traditional signature checks are performed
for the detection of known as all as the novel malware. An integrity check of
authorized Kernel modules is given which can prevent the installation of root kits
through the corrupted kernel modules. This approach is easy to maintain as only
change is to be made in the kernel of the system. Infrastructural security is focused
not the attacks against VM monitors.
Authors [5] in their research presented a review of various methods and tools
used for detection and prevention of intruders in Cloud computing. Four concepts
for the development of the CIDS identified are (a) automatic computing; (b) on-
cology; (c) risk management; and (d) fuzzy theory. The taxonomy gives two layers
functional layer and structural layer. The requirements identified for CIDPS on the
basis of review are (i) large-scale handling of multi tiered autonomous computing
and data processing environments; (ii) detection of variety of attacks with least
positive rates; (iii) super fast detection and prevention; (iv) self-adaptive automat-
ically; (v) CIDPS Scalability; (vi) deterministic; (vii) synchronization of autono-
mous CIDPS; (viii) resistance to compromise. A Cloud intrusion detection and
prevention system which meets all the requirements is considered to be good one.
Review of CIDS and Techniques of Detection of Malicious … 105
A framework which is an open source solution has been given in [12]. APIs and
interfaces are given which are used in the development of the security components
in a distributed manner and building of customized event correlation rules. The
framework consists of a collection of components which are organized in a hier-
archical manner. The three main components of the framework are probes, agents,
and security engines.
According to the three architectural layers, the security engines are organized in
a hierarchical manner. At the lowest layer, the raw security data collected by the
security engines. It can be offered by the Cloud provider as service which includes
IDSs; Log analyzers; and specific security mechanisms provided by the Cloud
platform. It is the responsibility of the Cloud provider, at the higher level, to enable
additional IDSs and attack them to independent VMs. The provider is able to
recognize the compromised virtual components of the clients by correlating the
information provided by the higher layer with the data collected by the lower layer.
Attack Evaluation Tree (AET) is used to represent the attack in a tree like structure.
The goal of the attacker is the root node while the access path is through the
offsprings.
Authors in the research [13] introduced three concrete MI attacks with a proof of
concept implementation based on existing tools. Three introduced MI attacks in this
paper are: memory scanning, template poisoning, and snapshot cracking.
Authors [14] described the differences between the traditional insider and insider
in the Cloud. The two types of insider threats identified in Cloud computing, viz.,
(a) at the Cloud providers end, and (b) at the Cloud clients end. Both have different
set of problems and area of attacks. The countermeasures of the insider threat in the
Cloud provider are:-At client side: Cryptographic techniques, geo-redundancy are
used; and At provider side: Separation of duties, logging, legal binding, and insider
detection models. The problems in various methods are: (a) IDS/IPS: in IaaS
host-based IDS can be used, (b) Separation of duties: As in Clouds, same person
has multiple roles, it is difficult to implement it in Clouds; (c) Attack origin iden-
tification: (i) In case of Clouds, the access is usually done by some remote com-
puter. So there are no physical evidence for the attack, only digital evidence like IP
etc can be used and (ii) In case of shared credentials, it is difficult to fix the
responsibly; (d) Single point of failure and data leakage: Access to console of
administrator can cause heavy loss of that, that without any sign of intrusion. The
countermeasures of the insider threat in the Cloud outsourcer are (a) At client side:
Log auditing, host-based IDS/IPS are used and (b) At provider side: Anomaly
detection, separation of duties, and multifactor authentication are used.
5 Comparative Study
The various techniques being used for the detection of the intruders revealed by
review are proactive forensics, graph-based analysis, honey pots (based on network
sensors), IDS (based on network sensors), system call analysis (host based user
Review of CIDS and Techniques of Detection of Malicious … 107
profiling), command sequences and windows usage events, file system, memory,
I/O, and hardware monitoring, metrics about user sophistication (Usage Anomalies),
insider threat specification language, knowledge graphs, customized minimal attack
trees, technological, social and educational and psychological parameters [15–21].
Selecting the best one is quite challenging. The characteristics desired in a CIDS
are (i) Scalable and distributed IDS for Clouds without the failure points,
(ii) Combination of Knowledge base and behavior base in order to detect the known
as well as the unknown attacks with reasonable false alarm rates, (iii) Avoid single
point of failure, (iv) the IDS should be protected from the attacks by isolating it,
(v) flexible architecture, (vi) take into consideration various service models and
requirements of user, (vii) dynamic policies as the security needs of each VM are
varied, (viii) reduction in data transfer cost by reducing the network bandwidth,
(ix) easy to adapt.
6 Future Scope
References
1. https://en.wikipedia.org/wiki/Cloud_computing
2. Forrester-2012, Cloud survey. http://www.bmc.com/industryanalysts/reports/forrester-2012-
cloud-survey.html (accessed May 2012)
3. Yusop ZM, Abawajy JH (2014) Analysis of insiders attack mitigation strategies. Procedia Soc
Behav Sci 129:581–591
4. Ren K, Wang C, Wang Q (2012) Security challenges for the public cloud. IEEE Internet
Comput 16(1):69–73. https://doi.org/10.1109/MIC.2012.14
5. Patel A, Taghavi M, Bakhtiyari K, Júnior JC (2013) An intrusion detection and prevention
system in cloud computing: a systematic review. J Netw Comput Appl 36(1):25–41
108 P. Oberoi and S. Mittal
6. Kholidy HA, Baiardi F (2012) CIDS: a framework for intrusion detection in cloud, systems.
In: 2012 ninth international conference on information technology—new Generations,
978-0-7695-4654-4/12 $26.00 © 2012 IEEE
7. Khorshed MT, Ali ABMS, Wasimi SA (2011) Monitoring insiders activities in cloud
computing using rule based learning. In: IEEE 10th international conference on trust, security
and privacy in computing and communications (TrustCom-2011), 16–18 Nov 2011
8. Gul I, Hussain M (2011) Distributed cloud intrusion detection model. Int J Adv Sci Technol
34
9. Schmidt M, Baumgartner L, Graubner P, Bock D, Freisleben B (2011) Malware detection and
kernel rootkit prevention in cloud computing environments. In: 19th Euromicro international
conference on parallel, distributed and network-based processing (PDP-2011), pp 603–610,
9–11 Feb 2011
10. Nkosi L, Tarwireyi P, Adigun M (2013) Insider threat detection model for the cloud.
978-1-4799-0808-0/13/$31.00 ©2013 IEEE
11. Yang S-F, Chen W-Y, Wang Y-T (2011) ICAS: an inter-VM IDS log cloud analysis system.
In: IEEE international conference on cloud computing and intelligence systems (CCIS-2011),
15–17 Sept 2011
12. Ficco M, Tasquier L, Aversa R (2013) Intrusion detection in cloud computing. In: 18th
international conference on P2P, parallel, grid, cloud and internet computing (3PGCIC-2013),
pp 276–283, 28–30 Oct 2013
13. Nguyen M-D, Chau N-T, Jung S, Jung S (2014) A demonstration of malicious insider attacks
inside cloud IaaS vendor. Int J Inf Educ Technol 4(6). https://doi.org/10.7763/ijiet.2014.v4.
455
14. Kandias M, Virvilis N, Gritzalis D (2013) The insider threat in cloud computing. In: Critical
information infrastructure security. Lecture notes in computer science, vol 6983. Springer,
Berlin, pp 93–103
15. Mehmood Y, Habiba U, Muhammad AS, Masood R (2013) Intrusion detection system in
cloud computing: challenges and opportunities. In: 2nd national conference on information
assurance (NCIA), pp 59–66, 978-1-4799-1288-9/13©2013 IEEE
16. Gupta S, Kumar P, Sardana A, Abraham A, A fingerprinting system calls approach for
intrusion detection in cloud environment. In: 4th international conference computational
aspects of social networks (CASoN-2012), published by IEEE, pp 309–314
17. Martinez-Moyano IJ, Rich E, Conrad S, Andersen DF, Stewart TR (2008) A behavioral
theory of insider threat risks: a system dynamics approach. ACM Trans Modeling Comput
Simul 18(2):7.1–7.27
18. Dileep Kumar G, Morarjee K (2014) Insider data theft detection using decoy and user
behavior profile. Int J Res Comput Appl Robot 2(2):51–55. ISSN: 2320-7345. www.ijrcar.in
19. Young WT, Goldberg HG, Memory A, Sartain JF, Senator TE (2013) Use of domain
knowledge to detect insider threats in computer activities. IEEE security and privacy
workshops
20. Wongthai W, Rocha F, Van Moorsel A (2013) Logging solutions to mitigate risks associated
with threats in infrastructure as a service cloud. In: International conference on cloud
computing and big data, pp 163–170
21. Claycomb WR, Nicoll A (2012) Insider threats to cloud computing directions for new
research challenges. In: Proceedings of the 2012 IEEE 36th annual computer software and
applications conference, pp 387–394. IEEE Computer Society, Washington, DC, USA ©2012
Review of CIDS and Techniques of Detection of Malicious … 109
Author Biographies
Dr. Sumit Mittal received his Ph.D. degree & Master’s degree
from Department of Computer Science & Applications,
Kurukshetra University, Kurukshetra. Presently, he is working
as Professor & Principal, M.M. Institute of Computer
Technology & Business Management, M.M. (Deemed to be
University), Mullana, Ambala, Haryana, India. He is a life
member of Computer Society of India and member of various
professional societies of India & Abroad. He is also a member of
various academics bodies of M.M. University, Mullana. He has
more than 35 publications in International/National Journals and
Conferences. His research area includes Cloud Computing,
Wireless communication and Distributed Environments.
DNA-Based Cryptography for Security
in Wireless Sensor Networks
Abstract Wireless sensor networks (WSNs) employ tiny nodes which accumulate
information in various applications and security is essential for sensor network
applications, such as military target movement, etc. To impart security and privacy
to tiny sensor nodes is challenging task due to the restricted capabilities of sensor
nodes in terms of computation, communication, memory/storage, and battery
power. This paper proposes DNA-based cryptography with the use of secure socket
layer. It is exploratory research of biological based cryptosystem. As in conven-
tional cryptography public/private key, pair is used for encryption/decryption
process, we herein propose a DNA-based system in which the key pair is generated
with the use of RSA algorithm and shared with SSL protocol. So, this proposed
system resolves some of the problems related with sensor nodes and here we attain
security in three stages, i.e., information security, computation security, and bio-
logical security.
1 Introduction
Security of data is the most significant concern over the last few decades.
Cryptography is one such technique widely manifested in an array of security
system. It is a promising approach toward security of data transmission over public
networks by encrypting the original data or messages. Cryptography has a close
relation to the disciplines of cryptology and cryptanalysis, which gives the hiding
text so that data cannot be read or modified by intruder. However, as security is the
main concern in every type of networks thus cryptography is usually associated
with disorganizing the original information (plaintext) into hiding information
(ciphertext) and then the complete process is reversed for getting the original
information. The whole task is done by cryptographer [1, 2]. In wireless sensor
networks, sensor nodes are associated with many problems like small storage
capacity, low battery power, etc., and are also prone to various undesirable attacks.
As many sensor networks are assigned crucial tasks such as in military applications,
etc. Thus, security becomes the main issue that requires the consideration at the
time of designing the sensor networks [3–6]. So a new emerging technique, i.e.,
DNA computing (Biological Computing) has profound applications toward
decoding the problem related with sensor nodes. In the past decade, DNA com-
puting has gained prominence, particularly in developing sustainable medium for
large-scale computation system [7]. The pioneering and revolutionary work on the
implementing DNA concept is in the solutions of applications like cryptography,
clustering, scheduling, forecasting, and even trying to apply this in signal and image
processing application. In this context, L. Adleman decoded the complex compu-
tational problem in the year of 1994 [8]. He reported a simple and straightforward
result that DNA has high storage and computational capability. This convergent and
versatile method presents broad substrate scope and excellent functionality toler-
ance as a result of which using DNA computing tries to solve the problem related
with sensor nodes. However, many DNA algorithms have also been proposed and
still the research is going on [9, 10]. Here, in this paper, DNA-based algorithm has
been proposed by using the concept of SSL. SSL is secure socket layer protocol
used for sharing the public key and digital certificates between the sensor nodes,
and for key generation, RSA algorithm is used which is very secure algorithm as the
key pair is generated with the help of two randomly selected prime numbers. Thus,
the problem with sensor nodes is anticipated to be solved in some sense.
In this paper, Sect. 2 provides a brief overview of DNA concept, Sect. 3
describes the proposed work, Sect. 4 gives an illustration with example of proposed
work and in Sect. 5 an analytical review is done considering the different aspects of
security.
2 Components of DNA
3 Proposed Work
In the proposed work, RSA system is used for key generation. RSA is one of the
first practical public-key cryptosystems and is widely used for secure communi-
cation. In RSA, the encryption key is public and differs from the decryption key
which is kept secret [13, 14]. In RSA, the key pair is generated with the help of two
prime numbers which are randomly selected.
SSL protocol is used for key sharing. SSL is secure socket layer protocol which
is basically used for securely exchanging the keys and digital certificate between
two sensor nodes [15, 16]. Thus, this protocol provides the functionality of con-
fidentiality and authenticity.
114 M. Poriye and S. Upadhyaya
3.2 Encryption
1. If two sensor nodes A and B want to communicate to each other then they
should have key pairs (public key and private key). The private key will be with
the sensor node. The sharing of public key among the nodes will be through
SSL.
2. Encryption Process: if node A sends a message to node B then first that par-
ticular message is converted into ASCII values. After that these ASCII values
are encrypted using public key of node B (which is shared by using SSL). Then
convert the encrypted message (mini cipher) to base-4 equivalent which is in the
form of 0, 1, 2, and 3. The next process is getting binary data from previous
values and finally converting these binary values into DNA base equivalent.
Table 1 shows the binary equivalent of nucleotide bases as taken in the
encryption process.
Both node A and node B authenticate each other by means of digital certificate
which are being assigned to every node before deploying in any environment. The
digital certificate is shared by both nodes A and B with secure socket layer protocol.
Thus, both sensor nodes confirm each other’s identity that they are not any
adversary.
1. First select the encryption/decryption key pair using RSA algorithm. These keys
are generated using two prime numbers that must be randomly selected.
2. Table 2 shows the encryption process.
To recover the original data, the intended recipient uses its private key and the
rest of the process is just reversed of the encryption. Thus, the description can be
done successfully. The proposed algorithm was implemented in Java for both
encryption and decryption process and the correctness of the proposed procedure
was verified.
Level of Security: In our proposed system, security has been achieved in the form
of information security (accomplished by RSA), computation security (in form of
binary data), and biological security (as DNA bases).
Security against attacks: Security is the main issue in wireless sensor networks
because of many attacks. Thus for getting the security, the seven security principles
[17, 18] should be abided by any network. Our proposed system promotes all these
seven security principles:
1. Confidentiality: It specifies that only sender and intended receiver can access the
original message. This is achieved by our proposed system as the public key is
shared via the secure socket layer. Thus, the encrypted message by public key of
another node (in terms of mini cipher) is completely secure. So the interception
attack is not possible here (Fig. 2).
2. Authentication: It helps to establish the proof of identities. By sharing of digital
certificate between sensor nodes no adversary node can pose the other node’s
identity. In the Fig. 3, node C (adversary node) sends a message to node B
posing as node A. This type of attack is fabrication which cannot be possible in
our system.
3. Integrity: Content of message cannot be manipulated before it reaches the
intended recipient. As our DNA cryptosystem acquires security in three stages.
So there is no chance of modification attack by any adversary node C (Fig. 4).
4. Non-repudiation: There may be the situation that node A sends a message to
node B and later on refuses that message is not sent by it. This situation cannot
116 M. Poriye and S. Upadhyaya
Fig. 2 Confidentiality
Fig. 3 Authentication
Fig. 4 Integrity
happen in our designed DNA cryptosystem because every node is having the
digital certificate of every other node. Thus, no one can deny something after
having done it (Fig. 5).
DNA-Based Cryptography for Security … 117
Fig. 5 Non-repudiation
5. Access Control, Availability and Signature: Access control means who can
access what. Because the two nodes verify their identities via SSL, they may not
be able to access any information without each other’s permission.
Node C may try to interrupt the two nodes A and B by some intentional action
but here it cannot be possible because node C should have their digital certificate to
make a connection with node B. So interruption attack is not possible.
6 Conclusion
References
1 Introduction
the particular record while publishing data. A lot of work has been done on data
anonymization technique in recent years. The aim of these techniques is to allow
retrieval of useful information from huge available data while protecting sensitive
information.
The prime aim of adversary is to obtain the information about sensitive attribute
that can be determined by linking various attributes of relation with each other in
the published data [2]. In a relation, there are a variety of attributes which are
classified as key attribute, quasi-attribute, sensitive attribute, and insensitive
attribute.
Key attributes which are meant for unique identification of records and these
attributes are generally removed while publishing the information. Examples of key
attributes are roll no, name, phone no, etc.
Quasi-attributes are those which are used for linkage of anonymized data set
with the aim to reach to sensitive information, e.g., birth date, age, gender, zip code,
etc.
Sensitive attributes are those which need not be disclosed, and the aim of an
attacker is to determine these, e.g., disease of person, salary, etc.
Insensitive attributes are those which can never be useful for attacker.
Any anonymization technique removes identifiers for the data set and aim is to
produce anonymized data set. The present techniques and methods restrict sensitive
information in published data but result in huge information loss or data distortion
that affects the efficiency of data mining process.
Meanwhile, method development for data publication is utmost important; thus,
the generated data remains useful, and sensitive information will be kept secret. The
emphasis of this paper is on various privacy-preserving models, anonymity oper-
ation, etc.
2 Related Works
Evolution of various methods for secrecy of sensitive data is the main intent of
privacy-preserving data publishing. In past years, research communities worked on
these issues and have proposed various approaches. There are various operations
which are applied to the data sets to make data anonymized. The operators which
are frequently used for the purpose of anonymization are operation of generaliza-
tion, suppression, perturbation, etc. The model for data anonymization uses these
approaches, and the idea of algorithm for anonymization is based on specific
anonymity operations.
1. Generalization is a process of substituting the substantive value against less
specific but semantically consistent value. It is achieved using the purpose of
hierarchy tree and associated with attribute of category quasi-identifiers. In
Fig. 1a the nodes PGT, TGT, or PRT are more specific as compared with node
Privacy Preservation Using Various Anonymity Models 121
teaching, whereas it can be seen the node School Employee is at the top of
hierarchal level with highest level of generalization.
The reverse process of generalization is called specification. Typically there are
five types of generalization with difference in their scope: Full-domain gener-
alization, subtree generalization, sibling generalization, cell generalization, and
multidimensional generalization.
Full-domain generalization: This type of operation has the rarest search area
among all rather it leads to highest data distortion. The basic idea behind this
approach is to generate all attribute values to a common level of certain hier-
archy [3] (e.g., Fig. 1a); if PGT, TGT generalized to teaching, then other
attributes such as clerk, peon, etc., are also to be generalized to nonteaching.
Subtree generalization is smaller in boundary than the above said. In this, all
nodes rooted at intermediate level are generalized to same level or none are
generalized [4, 5], for example. If PGT which is at the lowest level of hierarchy
is to be generalized to teaching, then any of its siblings, i.e., TGT or PRT, are
also to be generalized, whereas other nodes at the same level will not be
changed to their generalized level.
Sibling generalization: The boundary of this process is smaller than
full-domain and sub-tree generalization. In this, among all intermediate nodes,
some are to be generalized whereas nodes from rest set remain unchanged [3],
e.g., PGT is generalized to teaching whereas other TGT remains unchanged.
Commonly, in process of global recording, if one of the values of the root level
is generalized, the rest will also be generalized.
Cell generalization: This approach is a little bit different as compared to other
generalization techniques as this is meant only for a single record. In this
(b) Person
M F
122 D. Narula et al.
scheme of local recording, if one of the values at the root level is generalized,
the rest will remain unaffected [3, 6, 7], e.g., in Fig. 1a when node PGT gen-
eralizes to its parent node teaching, it can maintain the PGT value in data set.
Multidimensional generalization: This generalization emphasizes different
generalization for different combinations of values of quasi-identifiers [3, 7],
e.g., in Fig. 1a, b [PGT, M] can be generalized to [Teaching, Person], while
[PGT, F] generalized to [School Employee, F]. This scheme has less data dis-
tortion compared to full-domain generalization.
2. Suppression: This is another flavor of generalization. In this the original values
of attribute generally quasi-attribute is replaced by special symbol (e.g., #,*) and
makes the value of that attribute meaningless, e.g., in Fig. 2 the zip code of the
city attribute is suppressed up to different levels and due to suppression at
different levels will make data more anonymous.
3. Randomization: This works with an ability to anonymize the whole data set for
certain semantic preservation. In the existing privacy-preserving data mining
techniques, randomization is considered as of important technique. This method
always provides the knowledge discovery and a balance between privacy and
utility [8]. When the randomized data is transmitted to the recipient, recipient
would receive it using distribution reconstruction algorithm.
4. Bucketization: The basic aim is partitioning tuples of table into buckets and
further separate the quasi-identifiers with reference to sensitive attribute by
arbitrary permitting values for sensitive attribute in every bucket and set of
buckets with permuted sensitive attribute values as anonymized data [8]. The
process of bucketization is used to anonymize high-dimensional data.
3 Privacy Models
record shows the value of sensitive attribute as 3. That is, it contains three
multiple values so if one identifies the group, then unable to determine the value
of sensitive attribute as it is multiple in no’s. In our example, we have taken
salary as sensitive attribute and applied l-diversity on it.
l-diversity prevents the linkage attack but it suffers from the problem of
skewness attack and similarity attack. When attacker is able to determine the
value of sensitive attribute, moreover, this is based on frequency distribution.
But similarity attack occurs when the value of sensitive attribute are different but
semantically similar quasi-group.
3. t-closeness: l-diversity never puts a check on attribute linkage when the overall
distribution of sensitive attribute is skewed, let us consider an example by
assuming a table containing sensitive attribute as disease and 95% of the records
are suffering from Flu whereas only 5% are suffering from Cancer. Now let us
suppose a qid group exists in two exactly equal half. The first half represents
Flu, whereas the other equal half represents Cancer. Now, if a tuple owner is
Privacy Preservation Using Various Anonymity Models 125
falling in the second half of the qid group (which represents Cancer) then the
chances of threat are high, because it comes under 50% confidence level as
compared to 5% [9].
So for the sake of preventing skewness attack [14], t-closeness had been pro-
posed, which pertains that in any group sensitive attribute distribution on
identifying attribute be closer to the distribution of attribute in the overall table.
t-Closeness algorithm is based on the concept of earth mover distance function
which is used to measure the closeness between two distributions of sensitive
values and closeness to be within t. The result after applying t-closeness algo-
rithm is shown in Table 4. t-Closeness also suffers from the limitation of not
suitable selection of EMD function for numerical sensitive attributes and
degradation of data utility [9].
126 D. Narula et al.
In this section, ARX data anonymity software has been used for determining the
values of various parameters such as minimum, maximum information loss, lowest
re-identification risk, individuals affected by lowest risk, and average
re-identification risk. k-Anonymity algorithm, combination of k-anonymity and
l-diversity algorithm, and combination of k-anonymity algorithm with t-closeness
have been applied on three different data sets. Their explanation is specified as
under.
Explanation and Results of Data Set I
The first data set contains sno, name, pincode, sex, age, and disease. For the
purpose of anonymization, sno and name will be treated as identifying attributes so
Privacy Preservation Using Various Anonymity Models 127
k(3)
150
100
50
0
Min info Max info Lowest Individual Average
loss loss re affected re
iden fica on Risk iden fica on
Risk Risk
will not be disclosed. Attribute disease will be treated as sensitive. After applying
the experimental settings for various quasi-attributes and applying anonymization
algorithms as discussed in Fig. 3 shows the results for various parameters
graphically.
Explanation and Results of Data Set II
The second data set contains age, education, place, and occupation. For the purpose
of anonymization, attribute disease will be treated as sensitive. After applying the
experimental settings for various quasi-attributes and applying anonymization
algorithms as discussed, Fig. 4 shows the results for various parameters graphically.
Explanation and Results of Data Set III
The third data set contains age, designation, gender, zip code, and salary. For the
purpose of anonymization, attribute salary will be treated as sensitive, after
applying various hierarchical settings for quasi-attributes and applying
anonymization algorithms as discussed. Figure 5 shows the results for various
parameters graphically.
128 D. Narula et al.
250
200
150
100
50
0
Min info Max info Lowest Individual Average
loss loss re affected re
iden fica on Risk iden fica on
Risk Risk
160
140
120
100
80
60
40
20
0
Min info Max info Lowest Individual Average
loss loss re affected re
iden fica on Risk iden fica on
Risk Risk
Privacy Preservation Using Various Anonymity Models 129
5 Conclusion
Data sharing is an essential part for many organizations as data is spread at various
sites and available in different formats; data about the individual in its original form
must contain some attributes which need not be disclosed, i.e., sensitive in nature
and cannot be published directly. So, with the help of various privacy data models,
data should be anonymized and preserved. In this paper, various models for privacy
preserving have been discussed along with various linkage attacks and anonymity
operations. In the last segment of this paper, it has been analyzed that no one model
for anonymization gives a consistent result. Thus, one can say that not only the
technique but the data set also plays an important role as values of various
parameters varies from one data set to other. So, privacy-preserving methods needs
to be further researched. It can also be inferred that there is a need of hybrid
algorithm that can handle different scenarios.
References
1. Yang X, Ma T, Tang M, Tian W (2014) A survey of privacy preserving data publishing using
generalization and suppression. An Int J Appl Math Inf Sci 8(3):1103–1116
2. Byun J-W, Kamra A, Li N (2007) Effiecient k-anonymization using clutering techniues,
DASFAA 2007, LNCS 4443. Springer, Berlin, pp 188–200
3. LevFevre K, Dewitt DJ, Raghu R (2005) Incognito: efficient full-domain k-anonymity. In
Proceeding of ACM SIGMOD, pp 49–60, New York, 2005
4. Bayardo RJ (2005) Data privacy through optimal k-anonymization. In: International
conference on data engineering, pp 217–228, Washington, DC, USA, 2005
5. Fung, BCM, Wang K, Yu PS (2005) Top–down specification for information and privacy
preservation, In: Proceeding of 21th IEEE international conference on data engineering,
ICDE’05, pp 205–216, Tokyo, Japan 2005
6. Wong RCW, Li J, Fu AWC, Wang K (2006) (a, k)-Anonymity: an enhanced k-anonymity
model for privacy preserving data publishing. In: Proceeding of 12th international conference
on knowledge discovery and data mining, pp 754–759, Philadelphia, PA, 2006
7. Xu J, Wang W, Pei J, Wang X, Shi B, Fu AWC (2006) Utility-base anonymization using local
recoding. In: Proceedings of 12th international conference on knowledge discovery and data
mining, pp 785–790, Philadelphia, PA, USA, 2006
8. Mirashe MS, Hande KN (2015) Survey on efficient technique for annonymized microdata
preservation. Int J Emerg Dev 2(5):97–103, ISSN 2249-6149
9. Fung BCM, Wang, K, Fu AWC, Yu PS (2011) Introduction to privacy preserving data
publishing concepts and techniques. CRC Press, Taylor and Francis Group, New York, p 13,
ISBN 978-1-4200-9148-9
10. Sweeney L (2002) k-Anonymity: a model for protecting privacy. Int J Uncertan Fuzziness,
Knowl-Based Syst 10:557–570
11. Machanavajjhala A, Gehrke J, Kifer D, Venkitasubramaniam M (2006) l-Diversity: privacy
beyond k-anonymity. In: Proceedings of the 22nd IEEE international conference on data
engineering (ICDE), Atlanta, GA, 2006
130 D. Narula et al.
12. Ashoka K, Poornima B (2014) A survey of latest developments in privacy preserving data
publishing. Int J Adv Inf Sci Technol 32(32):1–10, ISSN 319:2682
13. Machanavajjjhala A, Kifer D, Gehrke J, Venkitasaubramaniam M (2007) l-Diversity: privacy
beyond k-anonymity. ACM Trans Knowl Discov Data 1(1): 1–57
14. Li N, Li T (2007) t-Closeness: privacy beyond k-anonymity and l-diversity. In: Proceedings of
21st IEEE international conference on data engineering ICDE), Istanbul, Turkey, April 2007
A Hybrid Security Mechanism Based
on DCT and Visual Cryptography
for Data Communication Networks
1 Introduction
3 The Proposal
Figure 1 depicts the whole procedure carried out at sender and receiver side. First,
the encryption is done on the secret textusing visual cryptography [15], and then, it
is hidden behind a cover image using DCT steganography [16] and the stego image
finally created is transmitted over the network. The reversed process is carried out at
the receiver side to extract the message again.
The different techniques used in the proposal have been described as follows.
In this proposal, the cryptographic technique used is secret key visual cryptography.
In this particular technique, the secret message is divided into n shares. Presence of
all the n shares can only help in extracting the message as n-1 shares cannot reveal
any information about the message. This technique is favoured over other
134 Y. Jain et al.
CommunicaƟon Channel
techniques because of its lower time complexity and this feature is a result of its
simple algorithm at receiver side. Only an XOR operation is needed at receiver side
to extract the message in this technique.
At the sender side, the input data string to be encrypted is represented by inp.
Share1 and Share2 are two encrypted data strings which are created by considering
value of input inp. In the end, these two data strings(Share1 and Share2) are
combined and transmitted for further operations. Share_Comb represents the
encrypted form of inp.
Len=length(inp);
for i=1 to Len
Share1(i)=randi([0,1]);
end
for i=1 to Len
if(a(i)==0)
Share2(i)=Share1(i);
else
Share2(i)=not(Share1(i));
end
end
for i=1 to (2*Len)
if(i<=Len)
Share_Comb(i)=Share1(i);
else
Share_Comb(i)=Share2(i-Len);
end
end
Transmit(Share_Comb);
A Hybrid Security Mechanism Based on DCT … 135
In the following algorithm, f represents the encrypted data. The input f is broken
into shares (Share1 and Share2), and then, exclusive-OR is performed on shares to
retrieve the original data. Here, msg stores the decrypted data.
Len= length(f);
for j=1 to Len
if(j<=Len/2)
Share1(j)=f(j);
else
Share2(j-(Len/2)) = f(j);
end
end
for i=1 to Len/2
msg(i)= xor(Share1(i),Share2(i));
end
return (msg);
INPUT: Cover image cover.jpg of WxH size and secret message stored in text.txt file
OUTPUT: A stego image
Step 1: img=imread('cover.jpg');
Step 2: temp=fopen('text.txt','r';)
Step 3: msg=fread(temp);
Step 4: msg_bin=dec2bin(msg);
Step 5: break img into 8x8 sized blocks of pixels
Step 6: Perform DCT on each block using dct2() function.
Step 7: Compress each block through quantization table.
Step 8: while (data left to embed ) do
getLSB of next DCT coefficient from cover image
get next LSB from msg_bin
replaceLSB of DCTcoefficient with message bit
end while.
Step 9: Inverse DCT is performed using idct2() function on each block.
Step 10: All the blocks are combined to form a stego image named stego_img.
136 Y. Jain et al.
4 Performance Metrics
For complete analysis of the proposed scheme, different parameters are used, which
are divided into following categories:
The parameters under this category measure the picture quality. Widely used
parameters are as follows.
The MAE represents the mean absolute error between the original image and the
stego image.
X
n X
m
jf ði; jÞ yði; jÞj
i¼1 j¼1
In the above formula, the mean absolute error is a mean value of the absolute
errors,. where f is the pixel value of original image and y is the true value of stego
image. Size of image is m n monochrome image. For coloured images, size of
image will be m n 3.
MSE stands for cumulative squared error between the stego image and the original
image. Lower value of MSE suggests lower error. It is defined by the relation given
below any m x n monochrome image.
A Hybrid Security Mechanism Based on DCT … 137
1 Xm1 X
n1
MSE ¼ ½Iði; jÞ Kði; jÞ2
mn i¼0 j¼0
It is defined as the ratio of peak square value of pixels by Mean Square Error
(MSE). It is expressed in decibel. The PSNR is defined as
MAXI2
PSNR ¼ 10: log10
MSE
where MAXI represents maximum value of pixel of the image. In the images with
pixel having 8 bits per sample, its value is 255.
The parameters under this category give the Qualitative measure of image and time
required to accomplish the process. Widely used parameters are as follows.
It is defined as the total processing time on receiver side or receiver and transmitter side.
In this paper, comparison of time consumption by all receiver side processes is taken.
The cover image may undergo change in pixel values during embedding operation as
a result of which the difference maybe observed in both the images. Hence, in order
to observe any change in visual quality, the qualitative visual analysis is helpful.
25
20
PSNR
15
10
0
192X192 256X256 320X320 384X384
ResoluƟon
2.00E+03
1.50E+03
MSE
1.00E+03
5.00E+02
0.00E+00
192X192 256X256 320X320 384X384
ResoluƟon
20
15
MAE
10
0
192X192 256X256 320X320 384X384
ResoluƟon
6
5
4
3
2
1
0
192X192 256X256 320X320 384X384
ResoluƟon
6 Conclusion
In this paper a secure and robust hybrid mechanism has been proposed which uses
DCT steganography and visual cryptography. Comparison results in Table 1 clearly
depict that the proposed technique is better than stand-alone DCT in terms of
various parameters as can be seen from the table below.
140 Y. Jain et al.
1. PSNR results clearly show that the proposed mechanism has a good picture
quality.
2. DCT has high time complexity. Since proposed mechanism uses visual cryp-
tography having low time complexity, therefore, the combination has moderate
time complexity.
3. Proposed mechanism provides a very good security.
References
1. William S (2003) Cryptography and network security: principles and practices. Pearsons
education, first Indian reprint
2. Preneel B Cryptographic algorithms: basic concepts and application to multimedia security.
Katholieke University, Belgium
3. Sadkhan SB (2004) Cryptography: current status and future trends. In: Proceedings of IEEE
international conference on information and communication technologies: from theory to
applications, Damascus, Syria, 19–23 Apr 2004, pp 417–418
4. Moskowitz I, Longdon G, Chang L (2000) A new paradigm hidden in Steganography. In:
Proceeding of the 2000 Workshop on new security paradigms, Ireland, pp 41–50
5. Bendor W, Gruhl D, Morimoto N, Lu A (1996) Techniques for data hiding. IBM Syst J 35(3
& 4)
6. Kharrazi M, Sencar HT, Memon N (2006) A performance study of common image
steganography and steganalysis techniques. J Electron Imaging 15:041104
7. Ker A (2004) Improved detection of LSB steganography in grayscale images. In: Proceedings
of information hiding workshop, vol 3200, Springer LNCS, pp 97–115
8. Raphael AJ, Sundaram V (2012) Cryptography and steganography—a survey. Int J Comput
Technol Appl 223–231
9. Jena D (2009) A novel visual cryptography scheme. In: IEEE international conference on
advanced computer control
10. Gokul M, Umeshbabu R, Vasudevan Shriram K, Karthik D (2012) Hybrid steganography
using visual cryptography and LSB encryption method. Int J Comput Appl 59:5–8
A Hybrid Security Mechanism Based on DCT … 141
11. Gupta S, Goyal A, Bhushan B (2012) Information hiding using least significant bit
steganography and cryptography. I.J. Modern Education and Computer Science 6:27–34
12. Nivedhita R, Meyyappan Dr T (2012) Image security using steganography and cryptographic
techniques. Int J Eng Trends Technol 3:366–371
13. Aung PP, Naing TM (2014) A novel secure combination technique of steganography and
cryptography. Int J Inf Technol Model Comput (IJITMC) 2:55–62
14. Shingote PN, Syed AH, Bhujpal PM (2014) Advanced Security using Cryptography and LSB
Matching Steganography. Int J Comput Electron Res 3:52–55
15. Nakajima M Extended use of visual cryptography for natural images, department of graphics
and computer sciences. Graduate School of Arts and Sciences, The University of Tokyo
16. Morkel T, Eloff JHP, Olivier MS (2005) An overview of image steganography. In: New
knowledge today conference, Sandton, pp 1–11
Author Biographies
Abstract Cloud computing delivers a broad range of services and resources like
computational power, storage, computational platforms, and applications to cloud
consumers through the Internet by on demand, pay-per-usage basics. With a
growing number of cloud service providers resorting to using and sharing resources
in the cloud environment, there is a necessity for protecting the data of various users
from unauthorized access of information between network and cloud. However, the
security and privacy of an open-ended, reasonably sharing of accessible resources is
still uncertainty and present a major complication for cloud consumers to accli-
matize interested in cloud environment. This manuscript initiates and deeply
examines the cloud security problem. This paper deals with the protection concern
that includes many of the cloud attacks, data integrity, data leakage, privacy,
confidentiality, vulnerabilities during sharing of resources, services, and informa-
tion. This method deals with securing the cloud information without data loss from
malicious users, hackers, and attackers of a real-time environment. This method
verifies user authentication and authorization management. It assures security on
the transmission of data, quality of service, and prevents vital information from
various active and passive attacks. This proficient method preserves the cloud
environment with better performance evaluation. Furthermore, security and privacy
analysis know the ability of the proposed method for cloud computing and extend
productive efficiency with safe cloud computing environments.
S. Srinivasan (&)
Research Development Center, Bharathiar University, Coimbatore, Tamilnadu, India
e-mail: [email protected]
S. Srinivasan
Department of M.C.A, K.C.G College of Technology, Chennai, Tamilnadu, India
K. Raja
Alpha College of Engineering, Chennai, Tamilnadu, India
e-mail: [email protected]
processing, and bandwidth. The benefits [4, 5] of cloud environment are shown in
Fig. 1.
In a cloud environment, security and privacy are shared between the cloud
service providers and consumers. The main problem of the cloud environment is a
large number of security threats, cloud attacks, hijacking of network information on
outsourcing of resources as well as business-critical process and data. Some
security issues in the cloud are data integrity, information confidentiality, data
leakage, vulnerability, and data intrusion. To ensure facts’ confidentially, infor-
mation integrity and availability, the cloud provider provides that at a minimum,
include the following:
• Cryptographic method to guarantee that the shared and global data centers
secured all information.
• Strong user access technique and user authentication methods to protect against
illegal contact to the data.
Cloud computing security [5] is a large set of security controls, strong policies,
recent technologies, and methods set to safeguard the data and applications of the
cloud environment. The rest of this chapter is ordered as follows: Sect. 2 discusses
cloud security risks and issues. Section 3 gives a detailed description of the pro-
posed exciting advanced dynamic authentic security method for cloud computing.
Section 4 shows the performance of experimental and their interpretation outcome.
Finally, Sect. 5 concludes the paper and further improvements.
The important problem on the cloud is data integrity. The confidential infor-
mation stored in the cloud storage may suffer from harm or damage during tran-
sition actions from or to the cloud service provider such as the recently assaulted
Linux’s servers [8, 9].
Recent approaches for preserving the privacy and secrecy of users information
stored in the cloud storages mainly include cryptographic encryption method
(HMAC). Proof of Retrievability (POR) [10] is cryptography method for remotely
checking the data integrity and confidential information stored on the cloud server.
Information confidentiality and validity of data can be assured through crypto-
graphic methods.
Bernd et al. [11] investigate vulnerabilities that are also a major security concern
in a cloud. The control issue is a matter of vulnerabilities, which explore two
examples as listed below:
• Virtualized networks offer inadequate network-based controls.
• Poor key management events
There are several areas of risks that could be identified, in which data and
information security was the rate by 91.7% [12] as exposed in Table 1.
Some of the major cloud computing issues and different attacks [13] are men-
tioned below:
• Data confidentiality
• Vulnerability
• Leakage and loss of control
• Insider threats and malicious attacks
• Data intrusion
• Service hijacking
• Availability
• Hypervisor viruses
• Injection attack
• Denial of service
• Man-in-the-middle attack
• IP spoofing
The advanced dynamic authentic security method for cloud computing assesses the
problem of security and privacy from the cloud architecture standpoint, cloud
delivery model, and cloud deployment model perspective. This method appraises
the problem of various attacks, data integrity, data leakage, information privacy,
confidentiality, and vulnerability during the sharing of resources in cloud com-
puting. It prevents scam, error, misuse of illegal access and rights in cloud com-
puting environment. Cloud environment allows authenticated and authorized users’
to access the confidential information and sharing of resources, which leads to
developing effective and efficient security framework in the cloud computing
environment. The need of filtering is enforced on the secure communication channel
between cloud service provider and cloud user. To construct the self-directed
security of protected cloud environment through an alliance with safety services
such as authentication, privacy, and confidentiality.
To make sure of allocation of distribution information and availability of service
by integrating cryptographic encryption method and protective sharing algorithm
with authentication method [14]. This method strengthens the security which helps
to control privileged user access and monitor activities of malicious users in a cloud
environment. The advanced dynamic authentic security model for cloud computing
is shown in Fig. 3.
This model applied a layered protective structure with different layers and
ensures information security, eliminate various attacks, vulnerable file in this cloud
environment.
148 S. Srinivasan and K. Raja
The authentication and audit control method layer enforces user identity and
validation mechanism through biometric authentic signature verification of users,
one-time password method via mobile-based access security, social security identity
card like Aadhaar card and separate user secret key. It controls and eliminates
malicious users, attachers, and hackers with the help of centralized logging
mechanism.
This dynamic security method also manages user access permission matrix
method and keeps track of user activities via logging mechanism. This protective
cloud computing environment provides an integrated extensive variety security
solution, which ensures information confidentiality and data integrity.
In this method, protective sharing algorithm together with cryptographic meth-
ods to develop protective and sharing of resources and information in cloud
computing environment [14]. The advanced dynamic authentic security model
adopts multidimensional security architecture in cloud computing environment.
paper proposes a strong dynamic security structure for cloud environment with
many safety features such as shielding sharing of resources through cryptographic
encryption mechanism with authentication techniques.
Future research on this work will include to develop a better auditing technique
with specific standard interfaces and protocols that can maintain high confiden-
tiality, security, integrity, and to meet more secure protected cloud environment.
References
1. Lin G (2012) Research on electronic data security strategy based on cloud computing. In:
2012 IEEE second international conference on consumer electronics, ISBN:
978-1-4577-1415-3, pp 1228–1231
2. Behl A, Behl K (2012) An analysis of cloud computing security issues. In: 2012 IEEE
proceedings world congress on information and communication technologies, ISBN:
978-1-4673-4805-8, pp 109–114
3. Uma S, Kanika L, Manish M (2011) Implementing digital signature with RSA encryption
algorithm to enhance the data security of cloud in cloud computing. In: 2010 IEEE 1st
international conference on parallel, distributed and grid computing (PDGC—2010), ISBN:
978-1-4244-7674-9, pp 211–216
4. Behl A (2011) Emerging security challenges in cloud computing. In: 2011 IEEE, ISBN:
978-1-4673-0126-8, pp 217–222
5. Chen D, Zhao H (2012) Data security and privacy protection issues in cloud computing. In:
2012 IEEE proceedings of international conference on computer science and electronics
engineering, ISBN: 978-0-7695-4647-6, pp 647–651
6. Garfinkel SL (2007) An evaluation of amazon’s grid computing services: EC2, S3, and SQS.
Technical Report TR-08-07, Computer Science Group, Harvard University, Citeseer, pp 1–15
7. Zhou M, Zhang R, Xie W, Qian W, Zhou A (2010) Security and privacy in cloud computing:
a survey. In: 2010 proceedings of sixth international conference on semantics, knowledge and
grids, ISBN: 978-0-7695-4189-1, pp 105–112, 2010 IEEE
8. Cachin C, Keidar I, Shraer A (2009) Trusting the cloud. ACM SIGACT News 40:81–86
9. RedHat. https://rhn.redhat.com/errata/RHSA-2008-0855.html
10. Xu J, Chang E-C, Towards Efficient Proofs of Retrievability in Cloud Storage, National
University of Singapore Department of Computer Science
11. Bernd G, Tobias (2011) Understanding cloud computing vulnerabilities. In: Co published by
IEEE computer and reliabilities societies, IEEE April 2011, pp 50–57
12. Carroll M, van der Merwe A, Kotz P (2011) Secure cloud computing benefits, risks and
controls. In: 2011 IEEE
13. Denz R, Taylor S (2013) A survey on securing the virtual cloud. J Cloud Comput Adv Syst
Appl 2:17
14. Srinivasan S, Raja K (2014) Security challenges in cloud computing. Int J Emerg Technol
Adv Eng 4(4):01–06, ISSN 2250–2459
15. Yu J, Lu P, Zhu Y, Xue G, Li M (2013) Towards secure multikeyword top-k retrieval over
encrypted cloud data. In: IEEE Trans Dependable Secure Comput 10(4):239–250. July/
August 2013
152 S. Srinivasan and K. Raja
Author Biographies
Abstract Cloud Computing is a vast technology with high economic benefits, with
low cost, many industries planning to store their information on cloud maintained
by a third party are Third-Party Storage provider (TSP). Sometimes curious or
malicious administrators may leak data in the TSP. CryptDB provides confiden-
tiality works by executing the user SQL queries about encrypted data using Onion
Encryption. Another threat is from cloud users, where they try to secure the systems
against external adversaries with a secured user login and secure end-to-end
encrypted connections. Though, the internal adversaries remain also the biggest
threat from this case. The proposed security method of applying Elliptic Curve
Diffie–Hellman Ephemeral in CP-ABE (Cipher Text Attribute-Based Encryption)
technique for Key Exchange Policy. CP-ABE is a fine-grained access control with
the policy of Attribute Authority (AA) to a user is having Secret Key (SK) based on
the set of character attributes. Elliptic Curve Diffie–Hellman Ephemeral is a
well-known technique in key exchange policy. Combining CP-ABE with Elliptic
Curve Diffie–Hellman Ephemeral is proposed for the ECDHE-ZeroVi’s framework.
Keywords Confidentiality Ciphetext Fine-grained access control
Cloud CryptDB Elliptic curves Diffie–Hellman ephemeral
K. K. Chennam (&)
Gitam University, Computer Science Engineering, Hyderabad, Telangana, India
e-mail: [email protected]
A. L. Muddana
Gitam University, Information Technology, Hyderabad, Telangana, India
e-mail: [email protected]
T. Munnavara
M.J.C.E.T, Information Technology, Hyderabad, Telangana, India
e-mail: [email protected]
1 Introduction
A legal agreement which stores data in public clouds, hybrid cloud, or community
cloud give the advantage of TSP and may change the data on demand, where the
cost of TSP is reduced control on data security. TSP is an untrusted environment.
Traditional access controls are not suitable for the TSP-hosted database. The major
target is to provide security in untrusted domains. Since different clients may have
different queries, the access to the data must be based on individual and for par-
ticular authorized clients. A basic common method to protect the data onto the
cloud environment is encryption before sending it from trusted environment.
Traditional data encryption requires a single key or pair of keys to encrypt or
decrypt the data. The database requires a key separately for each cell (means each
column in a row) using fine-grained access, to generate the key or to store the key
or to manage keys require a trusted key store.
Another approach is to decrypt all protected data cells with a single key or a key
pair. CryptDB has two threats: First threat, the adversary wants to get access to the
DBMS server and trying to snoop on private data onto the cloud. The DB
administrator is not a trusted one. To prevent this, Onion Encryption is giving
security on Cloud, where the data stored in Onion Encryption given in [1].
Second Threat is decrypting the data onto proxy servers the authenticated user
and transferring data there is no guarantee of internal adversaries or logged in users.
The framework ECDHE-ZeroVi’s propose the use of CP-ABE with Elliptic Curves
Diffie–Hellman Ephemeral to control access to data based on the data consumer
attributes. Data consumers with attributes that can satisfy the policy and has
authorized key to decrypt data (Fig. 1).
Two goals for these threats: the First one is executing a different number of
queries with a reduced amount of secured information revealed to the cloud server.
Another one data should be encrypted with perfect and strong crypto systems
with AES that avoids the cloud server by executing many SQL queries. To solve
this case practically, the server needs to access the decryption key, and the
adversary may access all data.
2 Related Work
CP-ABE embedded the access policy in cipher text [3] and CP-ABE secures that
the owner who encrypts the data have full control and have information about the
descriptors’ [3]. CP-ABE generates the secret key to the user to allow the data
access and data decryption when the subset of attributes are matched according to
the cipher policy. Still the CP-ABE and KP-ABE both are semi-trusted techniques.
Their proposed frame work includes the Multi-Authority ABE (MA-ABE) allows a
different attribute authority with a different data need to generate the secret keys for
user to decrypt the data based on different attribute sets.
CryptDB [1] is software that can access the encrypted data stored in cloud
database and each encrypted column in tables are stored using different encryption
algorithms. CryptDB works by executing SQL Queries over encrypted data on
cloud database server. The database administrator never gets decrypted data access.
But the CryptDB cannot provide one too many encryptions.
Our proposed framework approach builds on selected concepts from above and
provides data security for the data provider with better efficient access control and
overcome the problems from distributed access control. Elliptic Curve Crypto
Protocol [17]: An equation shown below is a solution for an elliptic curve.
y2 þ axy þ by ¼ x3 þ cx þ dx þ e ð1Þ
y2 þ a1 xy þ a3 y ¼ x3 þ a2 x2 þ a4 x þ a6 ð2Þ
K is and arbitrary constant and a1, a2, a3, a4, a6 are constants in K
To use elliptic curves in cryptography the main constraint is that the curve is
nonsingular.
y2 ¼ x3 þ 2x þ 5 ð4Þ
y2 ¼ x3 2x þ 1 ð5Þ
Equations (4) and (5) should meet now. An elliptic group the Galois Field Ep (a,
b) with x3 + ax + b mod p for 0 x p where a and b are positive integers, but
less than p where mod p substitutes as in Eq. (3).
Security in CryptDB Using Fine-Grained Access Controls … 157
A fixed prime numbers p with Galois Field Ep (a, b) group for a and b non-
variable constants.
Example: Let us assume the points P = (x1, y1) and Q = (x2,y2) in elliptic curve
group Ep (a, b) and O is the point at infinity.
Addition rules of Elliptic group Ep (a, b) are
(1) P + O = O + P = P
(2) If x2 = x1 and y2 = −y1 means P = (x1, y 1) and Q = (x2,
y2) = (x1 − y1) = −P then P + Q = O
(3) If Q 6¼ −P then P + Q = (x3, y3)
Where x3 = k − x1 − x2 mod p
y3 ¼ kðx1 x3 Þ y1 modp
The key is exchanged between Mr. X and Mr. Y with mixed colors. At the end
the secret key is generated to encrypt and decrypt. Let us assume that yellow color
is known by Mr. X and Mr. Y on agreement as shown in the figure.
Initially, RSA-RC4-SH4 were used, the client randomly selects a secret key and
encrypts and send it to server whoever have the secret key can decrypt the data at
any time. If the adversary got the secret key while sharing they can decrypt the data
ECDHE-ZeroVi means elliptic curve, Diffie–Hellman Ephemeral signed by RSA
key with zero visibility of adversaries with CP-ABE. Where Diffie–Hellman
Ephemeral means server generates a different public key for every new query
requested by the client even though the adversaries breaks one public key can get
decrypt of only one query related data and already the tables are secured Onion
Encryption by CryptDB. The Elliptic curve using (P-256) is almost equal with
3248-bit RSA key so the adversary will ever never can break the key.
Proposed Performance:
Let E is an elliptic curve over the finite filed Fp is given in the following form:
Y 2 ¼ X 2 þ aX þ b;
a; b F p andð4a3 þ 27b2 Þ 6¼ 0
As discussed above when Mr. X and Mr. Y agree on a key (Yellow Color) then
they first fix a finite field Fq, an elliptic curve E and base point B E (with high
order). To generate key first Mr. X chooses a random number a € Fq and keep it as
secret. Next Mr. X calculates aB € E is a public and shares with Mr. Y. Where he
also performs same steps and calculates bB and shares with Mr. X. Their secret
common key is P = abB E
Definition: An elliptic curve E over the field F is a smooth curve.
Y 2 þ a1 XY þ a3 Y ¼ X 3 þ a2 X 2 þ a4 X þ a6 ; ai F
EðFÞ is a set of points; ðx; yÞ F 2 satisfies point at infinity O:
Insert
Data
Data Data Cloud
Encrypted
Provider Encrypted Server
Data/
ECDHE CryptDB Proxy Query
– ZeroVi’s Server
Frame Work
Data
Provider TSP
Data Encrypted
consumer Data
Plain
Authentication Query/
Decrypted Proxy DB
KEY PAIR
DP DC
Diffie-Hellman Prime Random
ECDHE Key Exchange EC (P-256)
Number Generator
Encryptor
Query Encrypted
Data
DBMS Server
(Encrypted Ta-
Crypt DB Server bles)
CP-ABE and key pair (SK, PK) by ECDHE. And even though the adversaries
attack and break the key the adversaries can see the data which is related to query
because the data stored in Cloud Server is encrypted using Onion Encryption where
it gives high protection layers through Onion Encryption.
References
Author Biographies
Keywords Cloud computing security Confidentiality Integrity
Availability Trusted third party Information security Public-key
infrastructure Mitigating security threats RD model Scramble
Unscramble AES DES 3DES
1 Introduction
Computing has bloomed and expanded horizontally and vertically with lots of
innovations in the field. There has been a wavy graph with a nudge in the infor-
mation and communication age. The computing era initiated with mainframe
computers, traversing toward minicomputers to personal computers later and now
we have reached to the most noteworthy era, i.e., cloud computing era. Cloud
computing services are offered by identified Cloud Service Providers (CSP) across
the globe. The CSP is considered to be simply an extension from Internet Service
Provider (ISP) and Application Service Provider (ASP). At the very initial level,
ISP 1.0 was implemented where Internet was provided locally to the institute. Later
it got transformed to ISP 2.0 and ISP 3.0, where now the Internet services were
available globally and users were able to connect with telecommunications and
other service providers thru associated data centers. These further got evolved to
ASP (ISP 4.0), where not only the computing infrastructure but also specialized
applications were provided with a greater ease. But considering the problem of ASP
where only dedicated infrastructures were implemented, a newer version of ISP,
ISP 5.0, called CSP got evolved, where the computing infrastructure along with
applications are available on a shared basis. Cloud computing characterizes a model
transition—a transfer from product-based computing to a service orientated com-
puting [1]. The US National Institute of Standards and Technology has defined that
cloud computing is a technology that facilitates well-situated, need-based network
admission to a communal group of computing resources, e.g., servers, networks,
applications, and offerings that can be quickly given and free with negligible
management attempt or service supplier interaction [2]. The cloud encourages ease
of use and is collected of five necessary distinctiveness, three delivery models, and
four deployment models [3]. Services offered by cloud computing are supplied with
dynamism to the customer who owns their data on cloud. As per their demand and
their need, the customer can easily access the data from cloud, as it is shared across
the network, from any location at a very high speed. Apart from this, cloud also
provides a very greater space for each individual to store data. Its benefits like
multi-tenancy, i.e., sharing of resources at the network level, high scalability,
elasticity, and pay-as-you-go facility have made cloud computing a promising and
swiftly budding model.
These elementary taxonomies are usually known as the “SPI Model”, where it
stands for Software services, Platform services, and Infrastructure services
respectively [4].
and do not require integration of other systems also. The provider only does the
patching and necessary improvements. SaaS provides clients with network-based
access to the commercially available software which is kept centrally.
• Public Cloud
In this type of cloud, the service is obtainable by the general public or to any bigger
organization. The cloud provided, is usually under the ownership of an organization
that sells cloud services.
• Private Cloud
Over here, the cloud is open only to the single client, or a solitary organization.
These types of cloud can be supervised either by the organization or any third party.
It varies in 2 different forms, i.e., off-premise and on-premise. In off-premise, the
cloud used is generally managed by any third party while in on-premise; the cloud
is managed and owned by the organization that uses it.
• Community Cloud
In community cloud, the cloud is mutually used in numerous organizations and it
supports an explicit community that has shared anxieties. It could be under the
supervision of the organizations or of an intermediate party and may be situated
on-premise or off-premise.
168 D. H. Parekh and R. Sridaran
• Hybrid Cloud
This cloud is a combination of more than one cloud, i.e., private, public, or com-
munity which will exist as exclusive entities but are leaped mutually by the pro-
prietary skill that allows data and application transportability.
The cloud computing security and related work done on security issues with
either encryption technique or cryptography is mentioned in the first half. The next
part describes the needs for cloud computing security with respect to
Confidentiality, Infrastructure, and Availability (CIA) is discussed. This paper has
also proposed a model that shows the use of DES algorithm in encrypting and
decrypting process but involving scrambler and unscrambler. The proposed model
ensures all major needs that cloud security requires. It enhances integrity assurance
and confidentiality as well as the availability of data on cloud.
2 Related Work
It is observed that a large amount work has been carried out in the vicinity of cloud
security. A major portion of the work focuses on the reliability verification the
saved data in the cloud. Tangowan et al. [6] have depicted cloud computing security
anxieties that are specifically related to security of data and privacy-based guard
issues which has remained as a chief restraint for the implementation of services
provided by cloud computing. They have offered with summarizing but thorough
analysis on data security and privacy protection issues. But the disadvantage is that
it does not show any practical implementation of the security policy or mechanism.
Somani et al. [7] state that in cloud computing problems like data security, file
system, backup, and host security persists to a greater extent. They have projected a
notion of the digital signature with the use of RSA algorithms to encrypt the
sensitive data while shifting it over the network. This technique has tried to solve
the problem of authentication and confidentiality. But as observed, the problem of
integrity still persists.
Similarly, Rafique et al. [8] have shown a secure data transfer based on identity
in cloud using a method called Group Digital Signature (GDS). In this, a group
manager will commune with the service giver by using a secret key that will get
produced by the Diffie–Hillman key exchange algorithm. Group manager obtains
the member public key of all the users in the group. The user in the group sends the
data to the cloud server and will sign the message with the assigned (d, n) private
key. This message is acknowledged by the group manager who authenticates the
group member and then gathers the necessary detail and further attaches the secret
group id and sign and sends it to the cloud provider. Cloud provider will authen-
ticate the message and will allow the encrypted message to be stored in private
cloud. But as observed, one needs to trust the group manager, which might not be
feasible at every instance [5].
Mitigating Cloud Security Threats … 169
Moreover, Fernandes et al. [9], has shown in their paper that security related to
data in the cloud can be assured with the use of digital signature with help of
CFX_MF algorithms. In this digital signature is used for the verification and
non-repudiation of the message, where the uniqueness of sender and the reliability
of the message are preserved. According to the paper, the integrity check over the
cloud computing is performed by an intermediate party which inspects the data
from client and hauls out the request of unauthorized user. Some researchers do not
trust the third party as there is no guarantee of mutual and equal trust.
After surveying various papers, and flaws with the usage of encryption tech-
niques in the papers, this paper is aimed to focus on guaranteed cloud security
model. This model will use Data Encryption Standard (DES) algorithm with
scrambling and unscrambling of data and is also ensuring the mechanism to assure
data integrity and authentic data availability at the end once the encrypted data is
decrypted.
Securing data on the web involves recognizing exceptional threats and challenges
which necessarily has to be attended with greater impact by applying proper
countermeasures. Eventually, the required security services and controls are set up
with the typical systems engineering procedure in order to efficiently amalgamate
the defense controls with the information systems practical and equipped require-
ments, plus other significant system requirements like reliability, maintainability,
and supportability [10]. Usually, the architecture of cloud computing provides a
single data center for data storage and computation [11]. There can be various
security benefits in utilizing the cloud environment. But, a single malfunctioning
should not be alleged for any data loss. It generally is very difficult to track down
the security measures in a cloud environment. The current cloud service providers
have introduced and placed many complicated methods and trained staff for sus-
taining their systems. Due to this, there are various security benefits like data
centralization, data backup, incident response, logging, etc., available. Though it
shows the presence of many security features, cloud computing still addresses
major key security issues and challenges, like data segregation, usage of compro-
mised servers, certificates and auditing security, investigating an illegal undertak-
ing, and many more.
Cloud computing has become a most important development in IT. Enterprises
should acclimatize to the diversifications it brings to maximize the return on
investment. To assist organizations worldwide, International System Audit and
Control Association (ISACA) has identified critical issues which need operational
methods like effectively organizing risks, being transparent with the third party
about the enterprise policies, handling myriad regulations and adapting competently
[12]. In spite of several measures and steps for cloud security, the cloud has
exclusive features that involve endangering evaluation in fields like availability
170 D. H. Parekh and R. Sridaran
issues, data integrity, reliability problems, data recovery, and privacy and auditing,
as stated in Gartner [13].
Cloud computing, thus, as concluded, has a huge number of security issues and
challenges [14]. An elaborated record of security threats on the basis of the
deployment and service models of cloud computing is presented and discussed in
detail in [15]. Security, in general, to technology, is broadly standardized for
evaluation of data systems security, focusing on three central goals of CIA,
essentially known as, Confidentiality, Integrity, and Availability.
• Confidentiality
Confidentiality refers the access of restricted data only to authorized users and
ceasing access of such protected data from unauthorized users. Confidentiality aims
at authentication procedures like user-ids and passwords that solely recognize data
users and sustaining procedures that hamper each recognized user’s get access to
the system’s resources. But as there is augmented the quantity of parties, devices,
and applications occupied on cloud, the threat compromised data grow substantially
as the multiple access points come into existence. Such an increase in data usage
leads to problems with multi-tenancy, applications security, data remnants, and
privacy [16].
Cloud service providers usually are using a weak authentication mechanism that
involves username and password and the access controls, i.e., authorization, is at a
very coarse level, which results in significant security threats. To address these
security threats and to answer the cloud protection, in essence, there is a use of
encryption technique [17]. Encryption of data is carried out based on encryption
algorithm and is dependent on key strength. The encryption carried out even
depends on the cloud service providers; for example, EMC provides encryption
facility to the customer data while Amazon’s S3 does not provide any kind of
encryption to customer data but instead customer’s before uploading the data can
encrypt the data on their own.
The encryption of data for the purpose of providing confidentiality to customer
data primarily involves use of encryption algorithm. There are many encryption
algorithms present but not all are fashioned equal [18]. Cryptographically, many
algorithms are insufficient to provide the desired security. Algorithms that are
evaluated by formal standard bodies like NIST or informally by the cryptographic
community must be used. Next, the key length for encrypting data must be con-
sidered. It is essential to know that larger the key length, stronger is the encryption.
For the NIST-approved algorithms like 3 DES (Triple Data Encryption Standard)
minimum length should be of 112 bits, which will be shown in the proposed model.
• Integrity
Integrity is the next security aspect required for confidentiality. Integrity simply
means that consumer assets can be customized only by the authenticated users and
in an authorized way only. When it comes to data storage, maintaining data
Mitigating Cloud Security Threats … 171
integrity aspect is the obvious requirement. Data integrity ensures that no illicit or
illegal modification, deletion, or fabrication of data is allowed and originality of
data remains intact [19]. By keeping a check on the unauthorized access, organi-
zation attains greater confidentiality in terms of data integrity. Moreover, integrity
also helps in accountability of data modification, data deletion or any constructed
data, to find the potential source of such intrusion.
Data encryption is a solution for confidentiality but there should be a mechanism
to assure and verify the data that is decrypted by the recipient. This is taken care of
data integrity which uses message authentication codes tagged with the encrypted
data. These message authentication codes work as a hash function which will ensure
that the data that gets decrypted in the original sent message of the sender [20].
• Availability
The huge accessibility computing community has pursued a mantra that no par-
ticular source of failure should be observed, yet the administration of a cloud
examine by a lone company is, in fact, a distinct point of failure [21]. Availability
refers to every entity that comes when we talk about cloud. It targets the availability
of data in cloud, states the availability of the cloud service provider, system
availability and even talks about the availability of network level security mecha-
nism to ensure data security. Hence, availability is not only about data presence in
the cloud. Network is now getting highly congested and, therefore, need to assure
clients that the data will be available to them dynamically at any point of instance.
System availability involves the ability of system to continue with functioning in
a proper and accurate manner even when there is any kind of authority misbehaves
noticed. In spite of any security breach is identified, system should be able to carry
its operations as though normal. Cloud services show a severe reliance on the
resource infrastructures and network accessibility at all times [22].
Business critical applications generally rely on continuous and constant
delivery of services without a gap of any time. A simple service outage only for
few minutes can have a serious impact on the productivity of the enterprise. It can
also result in customer dissatisfaction and service-level disobedience. According
to the Cloud Computing Incidents Database (CCID) [23], which trails cloud
service outages, chief cloud service providers have undergone downtime ranging
from just minutes to hours. Moreover, relying on the rigorousness of the
occurrence and the extent of the exaggerated infrastructure, outages may involve
all or a few of clients. During a cloud service commotion, harmed clients will not
be in condition to contact the services and in a few cases can even experience
tainted presentation [24].
Apart from security concerns based on CIA, there are still many more other
concerns like privacy, data segregation, data storage, reliability, security, and data
leakage. But out of all, security is the major one where most of the researchers work
in the direction to secure cloud more day by day. To ensure the best security,
generally data transfer from host to server and vice versa happens with encryption
algorithms. Let us take a close look at few of the encryption algorithms.
172 D. H. Parekh and R. Sridaran
Earlier when data was stored on-premise, security measures were levied across the
institute, as the data used to be always on the traditional server residing in the
organization itself. But gradually, when we have started migrating on the cloud,
which is global, an essential security check to ensure data integrity, privacy, and
availability have become a major concern. To avoid the flaws, strong encryption
techniques are being implemented [25]. Below mentioned are kinds of encryption
algorithms used to ensure data security on cloud.
• RSA Algorithm:
This is the most commonly known algorithm, named after Rivest, Shamir, and
Adleman, the discoverer. It is a kind of asymmetric algorithm where an encryption
key is shared publicly to all but for encrypting a message, but the decryption key is
kept private and not publically. Moreover, RSA is a block cipher where each
message is charted in an integer. When used on the cloud, a cloud service provider
does the encryption of data, place the key publically and the user who accesses this
data from cloud, will decrypt it through a private key. RSA algorithm is found to be
secure only for the users, but doesn’t provide scalability and uses more of memory
space which is basic problems with RSA [26].
• DES Algorithm:
DES stands for Data Encryption Standard. It is a symmetric block cipher algorithm.
In this, data is encrypted in 64 bits of block size. Hence, 64 bits of data is input and
encrypted to 64 bits of cipher text. DES also ensures security at both the ends and is
scalable also. But it requires more memory space as compared to AES algorithm
[27].
• 3DES Algorithm:
3DES utilizes three occurrences of DES with different keys. It is deemed to be
secure because it needs operations enumerated to 2^112 to break it and none of the
recent technologies make it possible within the harmful duration of time. It is
inherently slow in case of of implementations, as it was premeditated to perform
on-chip rather than by chip [28]. Block diagram of Triple DES implementation is as
shown in Fig. 1 [29].
• AES Algorithm:
AES stands for Advanced Encryption Standard, and is a symmetric block cipher
kind of algorithm, used maximum nowadays. AES follows 128-bit key length for
encryption. In this type of algorithm, a data when is about to be stored on cloud by
the data generator, it is encrypted first and then this encrypted data is stored on
cloud. When any end users would like to use this data, the decryption takes place at
the data generator’s end and then only the users will be able to read data on their
side. AES is found to be highly scalable and is also providing security at both the
ends, i.e., users and the providers. Even the memory usage for AES kind of
encryption is found very low [30].
• Blowfish Algorithm:
Blowfish is a symmetric key cryptographic algorithm. It encrypts blocks of size 64
bits with a changeable length key of size 128–448 bits. Blowfish is suitable for
those applications where the key does not change frequently but remains constant
for a very long time. Blowfish is also secured for both the users and the providers,
and is also scalable. It is providing with good authenticity but is less used than AES
[31].
The encryption algorithms are very essential and provide a better mechanism to
secure data on cloud. As data security on cloud is the major concern, and as CIA are
very essential for cloud security, a model proposed below, known as RD Model, is
designed in such a way that it ensures all three very diligently.
5 Proposed RD Model
In this paper, an RD model is proposed which uses the first-level encryption fol-
lowed by scrambling and second-level encryption is carried out. As data security in
cloud computing is the most sensitive issue and is seeking utmost attention by
researchers, this paper aims at finding a solution for data security implementing
second-level encryption and scrambling of data. It also has depicted that the 2-level
data decryption ensures data confidentiality integrity and availability at a successful
results. The model is designed in MATLAB with implementation of 3DES algo-
rithm at present. It also shows the use of scrambling and unscrambling of data,
which ensures integrity and authenticity of transmitted data. In future, this model
will be demonstrated on Java platform to implement a real time model, so that data
security is guaranteed on cloud computing world without any doubts on the vul-
nerabilities of cloud.
References
1. Murugesan S (2011) Cloud computing gives emerging markets a lift. IT Pro, IEEE, pp 60–62
2. National Institute of Standards and Technology (2008) Guide for mapping types of
information and information systems to security categories. NIST 800-60
3. Hashizume K et al (2013) An analysis of security issues for cloud computing. J Internet Serv
Appl 4(1):1–13
176 D. H. Parekh and R. Sridaran
4. Zhu W, Luo C, Wang J, Li S (2011) Multimedia cloud computing. IEEE Signal Process Mag
59–69
5. Rimal BP, Choi E, Lumb I (2009) A taxanomy and survey of cloud computing. In: 2009 fifth
international joint conference on INC, IMS and IDC, IEEE, pp 44–51
6. Tangwongsan S, Itthisombat V (2014) A highly effective security model for privacy preserving
on cloud storage. Cloud Comput Intell Syst (CCIS). In: IEEE 3rd international conference
7. Somani U, Lakhani K, Mundra M (2010) Implementing digital signature with RSA
encryption algorithms to enhance the data security of cloud in cloud computing. IEEE
8. Rafique S et al (2015) Web application security vulnerabilities detection approaches: a systematic
mapping study. In: 16th IEEE/ACIS international conference on software engineering, artificial
intelligence, networking and parallel/distributed computing (SNPD). IEEE
9. Fernandes DAB et al (2014) Security issues in cloud environments: a survey. Int J Inf Secur
13(2):113–170
10. GroBauer B, Walloschek T, Stocker E (2011) Understanding cloud computing vulnerabilities.
IEEE, pp 50–57
11. Schneiderman R (2011) For cloud computing, the sky is the limit. IEEE Signal Process Mag
15–17
12. Heier H, Borgman HP, Bahli B (2012) Cloudrise: opportunities and challenges for IT
governance at the dawn of cloud computing. In: 45th Hawaii international conference on
system science (HICSS). IEEE
13. Gartner (2008) Assessing the security risks of cloud computing. Gartner
14. Parekh DH, Sridaran R (2013) An analysis of security challenges in cloud computing. In:
IJACSA
15. Cloud Security Alliance (2010) Top threats to cloud computing, Cloud Security Alliance
16. Harauz J, Kaufman LM, Potter B (2009) Data security in the world of cloud computing, IEEE,
pp 61–64
17. Aazam M et al (2014) Cloud of things: integrating internet of things and cloud computing and
the issues involved. 2014 11th international Bhurban conference on applied sciences and
technology (IBCAST)
18. Yu S, Wang C, Ren K, Lou W (2010) Achieving secure, scalable, and fine-grained data
access control in cloud computing
19. Yu S, Ren K, Lou W, Li J (2009) Defending against key abuse attacks in kp-abe enabled
broadcast systems, In: Proceedings of SECURECOMM’09
20. Wang C et al (2010) Privacy-preserving public auditing for data storage security in cloud
computing. In: Proceedings of INFOCOM. IEEE
21. Wang Q et al (2009) Enabling public verifiability and data dynamics for storage security in
cloud computing. In: Computer Security—ESORICS 2009. Springer, Berlin, pp 355–370
22. Armbrust M et al (2010) A view of cloud computing. Commun ACM 53(4):50–58
23. Zhang Q, Cheng L, Boutaba R (2010) Cloud computing: state-of-the-art and research
challenges. J Internet Serv Appl 1(1):7–18
24. Popović K (2010) Cloud computing security issues and challenges. In: MIPRO, proceedings
of the 33rd international convention. IEEE
25. Subashini S, Kavitha V (2011) A survey on security issues in service delivery models of cloud
computing. J Network Comput Appl. Elsevier, pp 1–11
26. Sun D et al (2011) Surveying and analyzing security, privacy and trust issues in cloud
computing environments. Proc Eng 15:2852–2856
27. Buyya R et al (2009) Cloud computing and emerging IT platforms: vision, hype, and reality
for delivering computing as the 5th utility. Fut Gener Comput Syst 25(6):599–616
28. Shao J, He Z (2004) High-speed implementation of 3DES encryption algorithm based on
FPGA. Mod Electron Technol
29. National Institute of Standard and Technology (1999) Data encryption standard (DES)[EB/
OL]. http://www.csrc.nist.gov/publications/fips/fips46-3/fips46-3.pdf
Mitigating Cloud Security Threats … 177
30. Sanaei Z et al Heterogeneity in mobile cloud computing: taxonomy and open challenges.
Commun Surv Tutorials 16(1):369–392
31. Xiao Z, Xiao Y (2013) Security and privacy in cloud computing. Commun Surv Tutorials 15
(2):843–859
Author Biographies
Prof. Disha H. Parekh, M.Phil., MCA, PGDBA (Human Resource), is presently an Assistant
Professor of Faculty of Computer Applications at Marwadi Education Foundation’s Group of
Institutions, Rajkot, Gujarat. She has completed her M.Phil. in Computer Science from Bharathiar
University and is at present pursuing Ph.D. in computer science on cloud computing. She did her
MCA from Ganpat University, Gujarat. She even completed PGDBA with a specialization in HR
from Symbiosis University. She has published 3 papers in the International Journal and has
presented 1 paper at National conference. She has attended many workshops and seminars. Her
areas of interest are Software Engineering and Web Technologies.
Dr. R. Sridaran, is currently the Dean, Faculty of Computer Applications, Marwadi Education
Foundation’s Group of Institutions, Rajkot, Gujarat. He did his postgraduation in Computer
Applications and Management. He was awarded Ph.D. in Computer Applications in 2010. Having
started his career as an Entrepreneur, he has offered his consultancy services to various service
sectors. He designed and delivered various training programs in the areas of IT and Management.
He has published 15 research papers in foremost Journals and Conferences and is currently guiding
five research scholars. He has got 22 years of academic experience and has served in principal
educational institutions at diverse capacities.
Analysis and Impact of Different
Mechanisms of Defending
Pass-the-Hash Attacks
Abstract Pass-the-hash attack has been around more than 18 years, and founded
roots of its first existence were around 1997. The reason for discussing it again now
is that it has come into forefront in recent times. And with the usage of Internet and
World Wide Web like never before and given the ever increasing use of the Internet
(2 billion users in 2011 with forecasts of another billion users coming online in the
recent 4 years) and excessive use of ubiquitous devices and computing it has
capabilities of affecting the most now. There are always certain machines which are
unpatched or those which can have easy to find opening and to get the different
privileged accesses through them, and now that cloud-based services are used and
major chunk of computers still working on Windows platform, which is most prone
to this kind of attack. It becomes a big threat for cloud services providers as well. In
this paper, we are trying to focus and understand pass-the-hash attack and also
discuss various pros and cons of some of the different approaches. Here, we present
results and approaches proposed by various researchers, and also address the
strengths and weaknesses of the solutions. Albeit extensive advancement has been
made, more research should be done to address this issue. We propose several
defense mechanisms here which are first of their kind and if implemented may
reduce the repercussions of the attack.
1 Introduction
More and more computers, BYOD (Bring Your Own Device), and clouds expand,
it invites huge number of people accessing Internet. By current numbers, it is
already in billions as per one of the statistics [15]. Before we talk on different
approaches or methods to defend against and reduce the mitigation of pass-the-hash
attack (PTH), “There are two types of companies today, those that have been
hacked and those that don’t know they’ve been hacked” [16]. Generally, passwords
are the most usually utilized security apparatus in the world today. Solid passwords
are the absolutely most imperative part of data security, and weak passwords are the
single biggest disappointments.
As the devices and strategies for credential theft and reuse attacks like the PTH
attack enhance, malignant clients are discovering it to be less demanding to
accomplish their objectives. The PTH attack is a standout among the most prevalent
sorts of qualification theft and reuse attack seen by Microsoft to date. Other per-
sonal information theft attacks incorporate key logging and plaintext secret key
capture, passing tickets, token mimic, and man-in-the-middle attacks. As mentioned
earlier also, there are always certain machines which are unpatched or those which
can have easy to find opening to get the different privileged accesses through them.
This does not mean that we should surrender to attackers or hackers.
But issues with the other types of attacks are also reason for hackers preferring
PTH attack. Password attacks, for example, watchword speculating or watchword
breaking is time-consuming attacks. Devices that make utilization of pre-computed
hashes diminish the time expected to acquire passwords incredibly. Nonetheless,
there is capacity cost and time utilization identified with the era of those precom-
piled tables; this is particularly genuine if the calculation used to create these
passwords is generally solid, and the passwords are mind boggling and long (more
noteworthy than 10 characters) [12].
In a PTH attack, the objective is to utilize the hash straightforwardly without
splitting it. This eradicates the need for password cracking or guessing algorithms
and procedures.
Before we can investigate the PTH attack, it is key to characterize a hash. For every
client and head account on a framework, the working framework stores the user-
name and a password with a specific end goal to perform authentication. On the
other hand, of putting away the password in clear content, the working framework
utilizes cryptographic hash capacities to make a hash esteem that it stores [20].
Analysis and Impact of Different Mechanisms … 181
At the point when a client tries to authenticate to the framework, the framework
takes the password info by the client, registers its hash esteem, and looks at the
figured hash against the put away hash. On the off chance that the hashes match, the
client is permitted access to the framework.
All the hashes are stored in computers Security Account Manager (SAM) file on
computer. This includes all the values such as individual accounts, administrator
accounts, or in that case any account details on the system.
In this attack, an attacker gets entrance to a client’s nearby regulatory hash and
afterward tries to utilize the hashes traded off from that framework to authenticate to
different frameworks on the system, conceivably obtaining entrance to extra hashes
along the way. The attacker then proceeds with this parallel development of trading
off distinctive frameworks inside of the system, increasing more hashes on each
bargained framework. A definitive objective is to obtain entrance to a special
domain account that can be utilized to get to discriminating servers and information
(Fig. 1).
2 Related Work
PTH is an attack that allows an attacker to use LM and NTLM hashes for
authentication remote (and local) station without knowing the password and
without breaking these hashes [14]. PTH attacks are no longer limited to only
certain functionalities [13]. They have evolved and are causing the problems in not
only client server environment but also are affecting the cloud networks as well.
Cloud computing is a service-based model, and hence lot of BYOD devices are
used to get the services which makes it vulnerable to the PTH form of attack.
Basic matrices can be utilized to evaluate the danger of presentation to potential
cybersecurity dangers, for example, PTH. It has further demonstrated that on a
fundamental level it is conceivable to process this metric continuously amid the
approval period of network security, hence giving network overseers the capacity to
design a network to minimize, or possibly wipe out introduction to these sorts of
attacks [7].
The attacks themselves are dependable, hard to uncover, and frequently utilize
exceptionally propelled hacking methods. Since they are propelled in nature,
delayed, and constant, the associations behind them need to have an abnormal state
of learning, propelled apparatuses, and skilled faculty to execute them. The attacks
are normally preformed in a few stages—observation, readiness, execution, getting
entrance, data social affair, and association upkeep. In each of the stages, attacks
can be recognized with diverse probabilities. There are a few approaches to expand
the level of security of an association keeping in mind the end goal to counter these
occurrences. Above all else, it is important to instruct clients and framework
chairmen on diverse assault vectors and furnish them with learning and assurance
so that the attacks are unsuccessful. Second, actualize strict security strategies,
which incorporate access control and confinements (to data or network), encrypting
so as to ensure data and introducing most recent security overhauls. At last, it is
conceivable to utilize programming IDS tools to identify such peculiarities [18].
There is also approach of defense in depth which can be used to reduce the threat to
overall system. Dynamic defenses must also be enabled, which change attack
surfaces to proactively defend a network [6].
There are certain set of rules, and requirements shall be met in order for PTH attack
to be successful. On a very basic level, a PTH attack depends on three principles
that are considered as follows:
1. The capacity to pick up administrative rights on the system putting away the
required hashes,
184 N. Jadeja and M. Vaghasia
There is no single activity an association can take to keep a PTH attack. Both
Microsoft and the NSA recommend the “Guard in-Depth” approach—they
encourage associations to confine and ensure neighborhood and domain regulatory
accounts through such procedures as making extraordinary local administrative
passwords and executing least privileged access [1]. Moreover, they both prescribe
confining inbound activity what s more, horizontal development on the system with
firewall rule (Fig. 4).
There are several mitigations which can be applied. We have named this as
Defense Mechanism 1, Defense Mechanism 2, and Defense Mechanism 3. They are
the defensive measures we can apply to the system in order to reduce the amount of
influence PTH can cause. After discussing this strategy, we have also suggested
additional measures to avoid compromising of security due to PTH attack, espe-
cially in a cloud-based environment where access of resources and management of
resources are done using various kinds of BYOD devices as well.
This mechanism is simple yet very effective. As mentioned earlier, we will have to
restrict the incoming traffic to our servers, by the use of firewalls and other tools.
Tools which act as layer 8 technologies in addition to 7 layers are also available
such as Cyberoam firewall. Many other similar firewalls cum monitoring devices
are available. These devices can play major role.
• What is the aim of this defense mechanism?
The aim is to limit the capacity of hacker/cracker from starting side long
development from a traded off workstation by blocking or scanning incoming
traffic.
• What methods will be deployed?
Confine every single stream of incoming traffic with all workstations aside from
those with expected movement beginning from trusted sources, for example,
helpdesk workstation, etc.
• Resultant Output
Even if the hacker/cracker finds access to any system, he cannot get access to the
any other system in the network or cloud.
• Technical changes if any:
Not required.
This mechanism works on restricting the high privileged account access if any.
• What is the aim of this defense mechanism?
This defense mechanism decreases the danger of admin level people by dividing
the authority levels among several different types.
• What methods will be deployed?
Restrict the access of servers and critical accounts from limited number of
systems. Also, avoid access of such critical accounts or servers from BYOD
devices. Dedicated systems and computers can be assigned to admins. Also, dif-
ferent tasks of admin can be divided into multilevel authorities. No configuration of
services or task scheduling should be done from other than the assigned systems.
Analysis and Impact of Different Mechanisms … 187
• Resultant Output.
No compromise will be made for the attacker to get an access as dedicated
systems are assigned and used. Also, multilevel authorities will reduce the risk.
• Technical changes if any
No technical changes are required just rephrasing and creating the various
policies related to authentication and authority.
This mechanism works on restricting the local accounts from getting administrative
privileges. This means on local workstation; also, the user will not login using
administrator account but will use with his personal login ids.
• What is the aim of this defense mechanism?
This defense mechanism confines the capacity of hacker/cracker to utilize nearby
local systems or their reciprocals for parallel development PTH assaults.
• What methods will be deployed?
Restrict the access of systems from remote location or remote devices. Windows
operating system above vista have this functionality inbuilt. Create different pass-
words for administrative accounts wherever necessary.
• Resultant Output
Even if the hacker/cracker gets the access, rights or passwords will not be able to
login into the network or travel in the network parallely.
• Technical changes if any
Use of security identifiers along with different privileges can be done.
Other than this, various other defense mechanisms can also be applied in order to
reduce the repercussion of the attack. The following are the more defense
mechanisms.
– Chunk the Pass:
As we know, there is not eventually any guard against this attack, but that will
not mean that we cannot do anything at all about it. It is not highly contrasting. It is
just a different variant of gray. As we have seen the system of handicapping
powerless password hashes conflict with APTs (progressed constant dangers),
notwithstanding when the intruders own device work shall right and dandy utilizing
more grounded password hashing function. The intruders or crackers did not realize
that the weaker hashing function was debilitated, so intruders surrendered trying
these attacks.
188 N. Jadeja and M. Vaghasia
The finest guard against these attacks is to keep the intruders from getting super
admin rights to use in any case. Tragically, that includes about each conventional
PC security barrier: slightest benefit client logons, antimalware programming, white
listing, firewalls, etc. Extracting the hashes from the memory can be made a bit
harder. Specifically, in Windows-based system, the hashing function of passwords
can be hauled out of memories for the accompanying logon sorts: intelligent,
cluster, administration, open, remote intuitive, and stored intuitive. That may appear
like each kind of logon you can consider, yet it does exclude system logons.
Likewise, log off procedure on regular basis expels the hash function of pass-
word from memory, in spite of the fact that it can be missing in place by appli-
cations and APIs, so you never know. One way to clear your password hashes out
of memory is to log off from system.
– Disjoin those ties:
We request users to utilize non-intelligent approaches to oversee PCs. Rather
than utilizing RDP (Remote Desktop Protocol), run with a support instrument that
permits you to interface with remote PCs. A large portion of the Microsoft
Management Console (MMC) apparatuses can be re-focused to remote access PCs.
Use PowerShell scripts rather—at any point in time does not ask for sending the
passwords.
A suggestion is disposing of all SuperAdmin from the system or at least reducing
the amount of privileges they have. In active directory setup of Windows system,
“delegation” can be used to give administrators simply the privileges they require
without giving those highest level privileges, for example, with administrator. None
of enterprises or domains prefers this to be done by a single admin handling these
high-level critical operations. Rather, utilize designation and hand out only the
authorizations and benefits important to handle the errands needed for those people.
In case of password hash getting stolen for one of the admins, the level of threat is
still very less comparatively, as the user is not super admin [4].
Other option is to work with OTPs, i.e., one-time passwords or very frequent
change in passwords. So even if the intruder does get the hash for the password, but
the time period to use will be reduced drastically. There are several different types
of tools that can help you in these both tasks. Also, a suggestion is to avoid the
recycle of the passwords on regular basis so the security remains intact [11].
Platform updates are regularly available and are generally automatically on. This
can be represented in summarized form as in Table 1.
Although PTH attacks proceed to represent a genuine danger for enterprises and
firms, by obtaining entrance to a client’s local administrative hash and traveling
through different workstations all through the network, an attacker can gain access
to a privileged domain account and use it to access critical servers and data. Still if
Analysis and Impact of Different Mechanisms … 189
Table 2 Platform updates and features to reduce effect of PTH attack [17]
Features Description Available Available Available Req. domain
on Win—7/ on Win—8/ on upgrade
Windows Windows Windows Windows
Server Server 8.1/Server Server 2012 R2
2008 R2 2012 2012 R2 domain
functional level
Remove LAN LAN manager legacy ✓ ✓ ✓
manager hashes and (reversibly
(LM) hashes and encrypted) plaintext
plaintext passwords are no longer
credentials from stored in LSASS
LSASS
Enforce New mechanisms have ✓ ✓ ✓
credential been implemented to
removal after eliminate session leaks
logoff in LSASS, thereby
preventing credentials
from remaining in
memory
Logon Use the new SIDs to ✓ ✓ ✓
restrictions with block network logon for
new well-known local users and groups by
security account type, regardless
identifiers (SIDs) of what the local
accounts are named
(continued)
190 N. Jadeja and M. Vaghasia
Table 2 (continued)
Features Description Available Available Available Req. domain
on Win—7/ on Win—8/ on upgrade
Windows Windows Windows Windows
Server Server 8.1/Server Server 2012 R2
2008 R2 2012 2012 R2 domain
functional level
Restricted admin The remote desktop ✓ ✓ ✓
mode for remote application and service
desktop have been updated to
connection support authentication
without providing
credentials to the remote
host
Protected users The new protected users ✓ ✓ ✓ ✓
security group security group enables
administrators to restrict
authentication to the
Kerberos protocol only
for group members
within a domain
Authentication New authentication ✓ ✓
policy and policies provide the
authentication ability to restrict account
policy silos authentication to specific
hosts and resources
proper care and right tools are used, then the defense mechanisms can be applied
and the repercussions are reduced [19]. The aim of this research was to get
acquainted with the attack “pass the hash” and working, then demonstrate it. This is
an attack that allows an attacker to misuse hashes credentials and usage errors in the
design of authentication protocols in order gaining access to high privileged
accounts. Our suggested defense mechanisms reduce the risk and also protect the
critical data and accounts.
As part of future work, we would like to test these mechanisms and approaches
in different environments. We would like to compare results with existing methods
and our approach for better accuracies and efficiency. Also, we would like to test
these mechanisms in cloud-based environment as well.
Here are some of the inbuilt and available features of Windows operating system
which can help reduce effect of the attack (Table 2).
References
3. Authentication Policies and Authentication Policy Silos published on 27 Nov 2013. Ref:
http://technet.microsoft.com/en-us/library/dn486813.aspx
4. Collecting Security Events Using Audit Collection Services in Operations Manager Ref:
http://technet.microsoft.com/en-us/library/hh212908.aspx
5. Framework for Improving Critical Infrastructure Cybersecurity by National Institute of
Standards and Technology February 12, 2014. Ref: http://www.nist.gov/cyberframework/
upload/cybersecurity-framework-021214.pdf
6. Groat S, Tront J, Marchany R (2012) Advancing the defense in depth model, In: 7th
international conference on system of systems engineering (SoSE), pp 285–290, 16–19 July
2012
7. Johnson JR, Hogan EA (2013) A graph analytic metric for mitigating advanced persistent
threat. In: IEEE international conference on intelligence and security informatics (ISI), vol no,
pp 129–133, 4–7 June 2013
8. McClure S, Scambray J, Kurtz G (2008) Hacking exposed 6: network security secrets &
solutions. McGraw-Hill, New York
9. Microsoft Security Compliance Manager originally published on 6 April 2010 and updated
January 28, 2013. Ref: http://technet.microsoft.com/en-us/library/cc677002.aspx
10. Microsoft Security Advisory 2871997 published on 9 Oct 2014. Ref: https://technet.
microsoft.com/en-us/library/security/2871997.aspx
11. Mitigating Pass -the-Hash (PtH) Attacks and Other Credential Theft Techniques by Microsoft
corporation, Published on 7 July 2014. http://download.microsoft.com/download/7/7/A/
77ABC5BD-8320-41AF-863C-6ECFB10CB4B9/Mitigating%20Pass-the-Hash%20(PtH)%
20Attacks%20and%20Other%20Cred-ential%20Theft%20Techniques_English.pdf
12. Pass the Hash attack, Microsoft Research as on August 12, 2015. Ref: http://www.microsoft.
com/PTH
13. Pass-The-Hash Toolkit for Windows Implementation & use by Hernan Ocha published on 29
Oct 2008. Ref: www.coresecurity.com/system/files/Ochoa_2008-Pass-The-Hash.pdf
14. Secrets of America’s Top Pen testers by Ed Skoudis Published in 2008. Ref: www.
inguardians.com/research/docs/Skoudis_pentestsecrets.pdf
15. Source: World Internet Stats: Usage and Population Statistics, 30 June 2010. Ref: http://www.
internetworldstats.com/stats.htm
16. The Year in Hacking, by the Numbers by NICOLE PERLROTH. Ref: http://bits.blogs.
nytimes.com/2013/04/22/the-year-in-hacking-by-the-numbers
17. TWC: Pass-the-Hash and Credential Theft Mitigation Architectures published by Mark
Simos, Nicholas DiCola published in TechEd North America on 14 May 2014. http://
channel9.msdn.com/Events/TechEd/NorthAmerica/2014/DCIM-B213
18. Vukalovic J, Delija D (2015) Advanced persistent threats—detection and defense. In: 38th
international convention on information and communication technology, electronics and
microelectronics (MIPRO), pp 1324–1330, 25–29 May 2015
19. What’s New in Remote Desktop Services in Windows Server updated on 28 May 2014. http://
technet.microsoft.com/en-us/library/dn283323.aspx
20. Why Crack When You Can Pass the Hash? By Chris Hummel published on 12 Oct 2009. Ref:
https://www.sans.org/reading-room/whitepapers/testing/ crack-pass-hash-33219
Data Security and Encryption
Technique for Cloud Storage
Abstract In the last few years, we have seen that cloud computing model has been
developed as a promising business model of the fastest growing IT sector. Most of
the IT companies, organizations, and educational institutes are now realizing that
they can put on fast access to daily used computer applications and significantly
boost up with infrastructure resources by simply moving to the cloud, at the very
negligible cost. But they are also worried about privacy and security of their data,
which is placed on the server of service providers. In this paper, we proposed a data
security and encryption technique to provide privacy and security to our dynamic
cloud data.
Keywords Secure cloud storage Cloud data security Cloud data encryption
1 Introduction
Today, we are living in an era, where technology plays an extremely vital role in
our daily lives and business. New technologies always bring greater ease and
convenience with them for us. Cloud computing is an endowed and evolving
technology in the field of network-based computing that takes place over the
Internet. It has become a well-known catchphrase nowadays. Cloud computing is a
model that facilitates the software developers to deploy their own applications,
S. Kumar (&)
Swami Vivekanand Subharti University, Meerut, Uttar Pradesh, India
e-mail: [email protected]
J. Shekhar
Computer Science Department, Swami Vivekanand Subharti University,
Meerut, Uttar Pradesh, India
e-mail: [email protected]
J. P. Singh
Shobhit University, Meerut, India
e-mail: [email protected]
Using the cloud storage system, users can easily store their personal or official data
(image, text, video, or audio) on the distributed cloud servers and this data will be
accessible anywhere, anytime through the use of the Internet. The availability,
security, and integrity of cloud data must be guaranteed by the service providers. As
data of an organization or an individual user is copied to several different sites to
minimize the data security threats from the hackers, therefore, an efficient and
optimized cloud data security technique plays the significant role in the cloud
computing. Many security techniques have been developed that give robust and
secure cloud storage and also give assurance of protecting the important data of
cloud users. As users frequently access or update their data from the different
locations using variety of devices like mobile, laptop, etc., in every aspects, security
to cloud data should be given. As a result, cloud storage is not only a third-party
warehouse but also there are many issues associated with them such as protecting
from unauthorized data access or modification and corruption of data, probably due
to the lack of server security. Many solutions have been given by the researchers so
that cloud storage can become trustworthy and users can use the cloud storage
service and store their data without any worry.
3 Related Works
In [3], authors have given a competent technique to illustrate the integrity of storage
data using hash index hierarchy and homomorphism verifiable response. They
present with the help of provable data possession concept, which supports a good
service and migration of data in the cloud environment.
In [4], authors proposed architecture with the approach for key exchange that
uses Diffie–Hellman key exchange along with the authentication step for each part
of the controller and the server instances. They suggest that data encryption using
RSA algorithm provides strong security over insecure medium. In their method,
they split the user data and encrypt, then send to the server instances for
computation.
In [5], authors highlight the various cloud data security issues and also give the
implementation of digital signature security technique using the elliptic curve P-192
in C language.
In [6], IBM discovers an encryption technique to improve cloud data security.
They also give the capability of spam filtering.
In [7], they propose one homomorphic encryption system that uses Residue
Number System (RNS), called HORNS. In that, they split the secret into multiple
sources to perform independent computations so that efficiency as well as security
can be increased.
196 S. Kumar et al.
4 Proposed Methodology
We have proposed a mechanism that uses a binary tree in that each node holds an
alphabet, number, and special character, and each link has a binary value 0 or 1.
Figure 1 depicts a binary tree with the values of numbers, special characters, and
alphabets. Depending upon their positions in the binary tree, a has 0, right sub-tree
node of a is c, that has 01, and left node of a will be b with having 00. Same as e
position is 001 and further elements of tree are depicted in the pair in Table 1.
a @ b c
0 00001 00 00
a b @ c
0 00 00001 01
Then, to make secure, put some bits like 1111 (Four times 1), after each bit.
In last step, apply the compression technique and send it to the cloud server for
storage.
7 Conclusion
We have developed a data security and encryption technique for providing security
to cloud users. Our proposed mechanism is unique and simple, but not easy for
hackers to crack because numbers increase the security while accessing the data. In
future, we will provide the cloud implementation of this technique, so that data of
cloud users can become more secure and cloud environment can become more
trustworthy.
References
Author Biography
Keywords Cloud computing Fine-grained access control Confidentiality
Attribute-based encryption Proxy re-encryption
1 Introduction
The main deployment models in cloud are public cloud, private cloud and hybrid
cloud. Public cloud is cheapest of all deployment models and is owned by third
party; however, they are highly insecure, for example, AWS. Private cloud is
owned by individual party and so is highly secure but at the same time they are
costliest, for example, Badaal Cloud. Hybrid cloud is owned partially by service
providers and partially by individual party and so are partially secured and is used
in mainly critical places like they are used in Union Bank of India.
Cloud computing several services are mainly categorized into three main types:
Infrastructure as a Service (IaaS), Platform as a Service (Paas), and Software as a
Service (SaaS). However, recently several types of service XaaS models are
defined; one of the such models is Data as a Service (DaaS) [6].
Among the several services, cloud storage service enables the owner of data to store
and share his important data with trusted clients which has freed the owner from
worry of storage and resource management. But at the same time since the owner
looses the physical control on stored data there are several security concerns related
Fine-Grained Access Control and Secured Data … 203
The main features to be achieved for securing data while outsourcing it on cloud are
as follows:
1. Data Confidentiality: Any unauthorized user or even the service provider must
not have an access to the data. Even if they steal the data, they must not be able
to decrypt it.
2. Fine-Grained Access Control: Each and every authorized user will be associ-
ated with some access rights. This enhances the efficiency and reliability in
system.
3. Improved Scalability: The system must be able to work efficiently with
increased number of users.
4. User Accountability: It should be maintained so that he can be charged
accordingly.
5. Efficient User Revocation: If the user is revoked, then the data owner need not
have to redistribute the keys to authorized user.
6. Efficient and Secure User Rejoin: If a revoked user rejoins with same or dif-
ferent access rights, then he must rejoin without affecting the system or users.
7. Collusion Resistant: There must be no collusion between the revoke user and
other authorized user or cloud service provider.
8. Ciphertext Size: The size of encrypted file must not be too big.
9. Support for Secured Query Processing: The encrypted query of authorized user
can be executed over an encrypted data and only the result of executed query
must be sent to authorized user.
10. Stateless Cloud: The cloud should not be in need to retain the state of revoked
and active users.
4 Related Work
For secured data sharing in cloud through CSP, many encryption schemes have
been introduced. The owner encrypts his data and sends it to third party called cloud
service provider. Along with encrypted data, owner also sends the access control
list specifying the authorization for accessing the attributes corresponding to users.
The cloud service provider converts the ciphertext of one authorized user to another
authorized user and provides it to him. In this way, data is securely shared among
authorized users using concept called fine-grained access control in order to limit
the access of encrypted data in cloud.
Fine-Grained Access Control and Secured Data … 205
In the traditional approach, if the owner wants to share some messages with others,
he should know public key authorized user in order to encrypt the data.
Identity-based encryption has changed the concept and allowed the public key to be
of random string, e.g., email id of recipient. One of the main issues arises from
sharing keys is user revocation where a user is needed to be revoked from accessing
his data. The usual solution followed by owners is to re-encrypt the whole dataset
with new generated key and redistribute the re-encrypted data to all authorized
users.
Sahai and waters presented attribute-based encryption in 2005 [32] for secured
data sharing based on the concept of public-key cryptography in which authorized
users are allowed to decrypt the data only if they satisfy certain attributes. The main
feature of this approach is that it is collusion resistant but since it uses access of
monotonic attributes in order to control users access, it is restricted in real envi-
ronment. Attribute-Based Encryption (ABE) was further classified as KP-ABE and
CP-ABE.
In 2006, Goyal [33] proposed KP-ABE in which users’ private key is used to
store access control policy and encrypted data stores additional attributes. An
authorized user can decrypt data if the access policy defined in users’ private key
satisfies attribute of data. However, the main issue with KP-ABE is owner (one who
has encrypted data) cannot take a decision on who can decrypt the data.
In 2007, Bethencourt et al. [31] introduced CP-ABE in which the access policy
is stored with encrypted data and attributes are stored in users’ secret key; as a
result, the user can access only the attributes associated with his private key. The
concept supports access control in real-time environment; however, it requires
flexibility and efficiency and its decryption key only supports user attributes that are
logically organized as a single set; as a result, user has to use a combination of all
attributes. To overcome this problem, ciphertext-policy attribute-set-based
encryption is introduced. It organizes user attributes into a recursive set-based
structure and user combines these attributes dynamically in order to satisfy a policy
without sacrificing the flexibility. The main challenge is allowing users to combine
attributes dynamically within a given key and avoiding collusion at the same time.
Earlier, ABE was based on monotonic access structure. Ostrovsky et al. in 2007
[34] proposed ABE that supports non-monotonic formulas on access policies to
express any access formula. Tang et al. in 2008 [35] put forward verifiable ABE.
Muller in 2009 [36] proposed an extension of CPABE, DABE (Distributed
Attribute-Based Encryption) that supports random number of parties to maintain the
attributes along with their corresponding secret keys; however, the access policy
has to be in DNF form.
Boneh and Franklin [37] proposed an identity-based encryption scheme, in
which data is encrypted using a random string as the key and for decryption; a
decryption key is mapped to the random encryption key-by-key authority.
206 N. Agarwal et al.
The main security concern while sharing the data using cloud is to prevent it from
semi-trusted cloud service providers. In order to maintain confidentiality, several
proxy re-encryption techniques are available. Proxy encryption is a primitive which
helps in translating ciphertext from one encryption form to another encryption form
without any information leaked to third party or cloud service provider. Application
of proxy re-encryption is sharing public health records online, social media, and
email forwarding.
The scheme proposed by Tang [41] enables owner to categorize ciphertext into
subsets and uses one key pair in order to simplify key management problem. These
subsets are re-encrypted to ciphertext using public key of specified authorized user.
The main advantage of this scheme is that every authorized user can use a particular
proxy.
It was introduced by Ateniese [42] in 2009 under this scheme that it is impossible
for proxy server to identify the recipient of the message.
Identity-based proposed by Shamir [43] uses string of arbitrary length such as email
id for creating public key of authorized users. The proxy server will translate the
ciphertext of Alice to ciphertext of Bob without being able to retrieve any
information.
Table 1 Comparison of attribute-based encryption
Techniques ABE KP-ABE CP-ABE IBE HABE DABE MA-ABE
Fine-grained access control Low High if there is Avg, high if there Avg Good Good Good
re-encryption, low is re-encryption
Efficiency Avg High for broadcast Avg Low Flexible Avg High
type system average
Fine-Grained Access Control and Secured Data …
Under this scheme, the owner specifies the conditions along with ciphertext and the
proxy can transform the ciphertext of data owner to encrypted form of recipient if
and only if ciphertext satisfies the condition specified by the owner. This scheme is
not sufficient to implement fine-grained access control [44].
The scheme introduced by Liu [45] has achieved user revocation and fine-grained
access control in the absence of data owner. In it, each user is associated with time
period for validity of user access rights so if he wants to access the data he needs to
have the access rights on attributes as well as access time must satisfy the validity.
Major limitation in it is for a user; the access time for all the attributes is same.
This scheme integrates encrypting, encoding, and forwarding [46] and exhibits
homomorphism, proxy re-encryption, and threshold decryption properties.
Homomorphism states that for ciphertexts c1 and c2 defined on plain text p1 and
p2, one can use c1 and c2 to obtain ciphertext on the plain text p1 p2 or p1 + p2.
Proxy re-encryption allows encrypted form of data of user1 to be transformed into
encrypted for another user without any information leaked to third party. Threshold
encryption lets the private keys to be divided into several pieces and distributed to
clients and all clients must together decrypt the file.
The scheme also fails to handle collusion between revoked and authorized user and
revoked user and untrusty cloud service provider.
Bharath et al. [50] proposed framework using proxy re-encryption and additive
homomorphic encryption in order to give a solution. He has implemented the
concept of federation of clouds in order to prevent collusion. However, there is a
limitation in their work that they have assumed that if revoked user colludes with
authorized user; then, the revoked user shares information available to authorized
user only (Figure 2).
One of the problems while outsourcing the data to cloud is that the query must be
executed and output should be given to only authorized users who have initiated the
query. While the query is being sent and processed and output generated, the
process should not be accessible to any unauthorized user or cloud service provider.
Boneh et al. [51] have presented a general framework for analyzing security of
searching on encrypted data systems. Under this framework, they have constructed
public-key systems that support comparison queries on encrypted data as well as
more general queries such as subset queries.
Hakan et al. [52] have introduced an algebraic framework in which they have
deployed coarse index which allows query to be partially executed on encrypted
data at providers end and then decrypted at client end and remaining query
executes.
210 N. Agarwal et al.
Hore et al. [53] have developed a bucketization procedure for answering mul-
tidimensional range queries on multidimensional data and allow the data owner to
control the tradeoff between risk and cost.
Wang et al. [54] have ensured data confidentiality both at storage and at access
time and also supports different queries and data updates.
Sharing data on cloud is widely accepted and is increasing rapidly. The data owners
are interested in outsourcing the data on cloud in order to avoid storage manage-
ment and capital expenditure in infrastructure but there are several issues associated
with it and one of the major issues is confidentiality and security. In this paper, we
have discussed on how to increase confidentiality and maintain privacy and security
while sharing the critical data through third party named cloud service providers.
We have explained encryption technique like ABE and PRE, when combined
altogether enable us to share the data securely maintaining confidentiality along
with fine-grained access control. However, the information can be leaked if there
exists collusion between cloud service provider and revoked user or between
authorized user and revoked users. Our proposed approach is to implement
multi-party computation-based homomorphic threshold cryptosystem under this
approach; private key of authorized user will be shared among n number of clouds
and the secret can be revealed if x out of total n participants work together. This
approach will prevent the data as the revoked user cannot collude with x number of
users altogether.
References
7. Kantarcioglu M, Clifton C (2005) Security issues in querying encrypted data. In: Proceedings
of the 19th annual working conference on data and applications security, DBSEC’05.
Springer, Berlin, pp 325–337
8. Cantor S, Sigaba JM, Philpott R, Maler E (2005) Metadata for the OASIS security assertion
markup language (SAML) v2.0”, copyright © OASIS open
9. Chow R, Golle P, Jakobsson M, Shi E, Staddon J, Masuoka R, Molina J (2009) Controlling
data in the cloud: Outsourcing computation without outsourcing control. In: Proceedings of
the 2009 ACM workshop on cloud computing security (CCSW), pp 85–90
10. Dahbur K, Mohammad B, Tarakji AB (2011) Security issues in cloud computing: a survey of
risks, threats and vulnerabilities. Int J Cloud Appl Comput (IJCAC) 1
11. Dhage SN, Meshram BB, Rawat R, Padawe S, Paingaokar M, Misra A (2011) Intrusion
detection system in cloud computing environment. In: Proceedings of the international
conference & workshop on emerging trends in technology, ICWET’11, pp 235–239
12. Kandukuri B, Paturi V, Rakshit A (2009) Cloud security issues. In: IEEE International
conference on services computing, pp 517–520
13. Singh G, Sharma A, Lehal MS (2011) Security apprehensions in different regions of cloud
captious grounds. Int J Network Secur Its Appl (IJNSA) 3
14. Brodkin J. Gartner: seven cloud-computing security risks. http://www.infoworld.com/d/
security-central/gartner-seven-cloud-computing-security-risks-853
15. Ristenpart T, Tromer E, Shacham H, Savage S (2009) Hey, you, get off of my cloud:
exploring information leakage in third—party compute clouds. In: Proceedings of the 16th
ACM conference on computer and communications security, CCS’09. ACM, New York,
pp 199–212
16. Wang C, Wang Q, Ren K, Lou W (2009) Ensuring data storage security in cloud computing.
In: International workshop on quality of service, pp 1 –9, July 2009
17. Salesforce.com. warns customers of phishing scam. http://www.pcworld.com/article/139353/
article.html
18. Yan L, Rong C, Zhao G (2009) Strengthen cloud computing security with federal identity
management using ierarchical identity-based cryptography. In: Proceedings of the 1st
international conference on cloud computing, CLOUDCOM’09. Springer, Berlin, pp 167–177
19. Lu R, Lin X, Liang X, Shen XS (2010) Secure provenance: the essential of bread and butter of
data forensics in cloud computing. In: Proceedings of the 5th ACM symposium on
information, computer and communications security, ASIACCS’10. ACM, New York
20. Lin D, Squicciarini A (2010) Data protection models for service provisioning in the cloud. In:
Proceeding of the 15th ACM symposium on access control models and technologies,
SACMAT’10, pp 183–192
21. Nyre AA, Jaatun M (2009) Privacy in a semantic cloud: whats trust got to do with it? In:
Cloud computing, volume 5931 of lecture notes in computer science. Springer, Berlin,
pp 107–118
22. Pearson S, Shen Y, Mowbray M (2009) A privacy manager for cloud computing. In:
Proceedings of the 1st international conference on cloud computing, CLOUDCOM’09.
Springer, Berlin, pp 90–106
23. Thuraisingham B, Khadilkar V, Gupta A, Kantarcioglu M, Khan L (2010) Secure data storage
and retrieval in the cloud. In: Collaborative computing: networking, applications and
worksharing (collaboratecom), pp 1–8, Oct 2010
24. Uemura T, Dohi T, Kaio N (2009) Availability analysis of a scalable intrusion tolerant
architecture with two detection modes. In: Proceedings of the 1st international conference on
cloud computing, CLOUDCOM’09. Springer, Berlin, pp 178–189
25. A. S. A. event. July 20, 2008. http://status.aws.amazon.com/s3-0080720.html
26. Takabi H, Joshi J, Ahn G (2010) Security and privacy challenges in cloud computing
environments. IEEE Secur Privacy 8(6):24–31
27. Jansen W, Grance T (2011) Draft special publication 800-144: guidelines on security and
privacy in public cloud computing. National Institute of Standards and Technology, U.S.
Department of Commerce
212 N. Agarwal et al.
28. Andrei T (2009) Cloud computing challenges and related security issues
29. Agudo I, Nuez D, Giammatteo G, Rizomiliotis P, Lambrinoudakis C (2011) Cryptography
goes to the cloud. in secure and trust computing, data management, and applications, vol 187
of communications in computer and information science. Springer, Berlin, pp 190–197
30. Santos N, Gummadi KP, Rodrigues R (2009) Towards trusted cloud computing. In:
Proceedings of the 2009 conference on hot topics in cloud computing, HOTCLOUD’09,
Berkeley, CA, USA. Usenix Association
31. Bethencourt J, Sahai A, Waters B (2007 )Ciphertext-policy attribute-based encryption. In:
Proceedings of IEEE symposium on security and privacy
32. Sahai A, Waters B (2005) Fuzzy identity-based encryption. In: Cramer R (ed) Advances in
cryptology—EUROCRYPT 2005. Springer, Berlin, pp 457–473
33. Goyal V, Pandy O, Sahai A, Waters B (2006) Attribute based encryption for fine-grained
access control of encrypted data. In: Proceedings of ACM computer and communications
security conference, CCS’06
34. Ostrovsky R, Sahai A, Waters B (2007) Attribute-based encryption with non-monotonic
access structures. In: Proceeding of ACM conference on computer and communications
security, pp 195–203
35. Tang Q, Ji D (2010) Verifiable attribute-based encryption. Int J Network Secur 10(2):114–120
36. Müller S, Katzenbeisser S, Eckert C (2009) Distributed attribute-based encryption. In:
Proceedings of 11th international conference on information security and cryptology (ICISC
08), pp 20–36
37. Boneh D, Franklin MK (2003) Identity-based encryption from the weil pairing. SIAM J
Comput 32(3):586–615
38. Boneh D, Boyen X, Goh E-J (2005) Hierarchical identity based encryption with constant size
ciphertext. In: Cramer R (ed) Eurocrypt, volume 3494 of lecture notes in computer science.
Springer, Berlin, pp 440–456
39. Wang G, Liu Q, Wu J (2010) Hierarhical attribute-based encryption for fine-grained access
control in cloud storage services. In: Proceedings of ACM conference on computer and
communications security, CCS’ 10
40. Wan Z, Liu J, Deng RH (2012) HASBE: a hierarchical attribute-based solution for flexible
and scalable access control in cloud computing. IEEE Trans Inf Forensics Secur 7(2):743–754
41. Tang Q (2008) Type-based proxy re-encryption and its construction. In: Proceedings of ninth
international conference on cryptology in India, pp 130–144
42. Ateniese G, Benson K, Hohenberger S (2009) Key-private proxy re-encryption. In:
Proceedings topics in cryptology, pp 279–294
43. Shamir A (1984) Identity-based cryptosystems and signatures schemes. Adv Cryptol 47–53
44. Libert B, Vergnaud D (2008) Tracing malicious proxies in proxy re-encryption. In:
Proceedings of PAIRING’08. LNCS 5209. Springer, Berlin, pp 332–353
45. Liu Q, Wang G, Wu J (2012) Time-based proxy re-encryption scheme for secure data sharing
in a cloud environment. Information sciences (in press)
46. Asharov G, Jain A, Lopez-Alt A, Tromer E, Vaikuntanathan V, Wichs D (2012) Multiparty
computation with low communication, computation and interaction via threshold FHE. In:
Proceeding of eurocrypt’12. Springer, Berlin, pp 483–501
47. Yu S, Wang C, Ren K, Lou W (2010) Achieving secure, scalable, and fine-grained data
access control in cloud computing. In: Proceedings of IEEE international conference on
computer communications, INFOCOM’10
48. Blaze M, Bleumer G, Strauss M (1998) Divertible protocols and atomic proxy cryptography.
In: Proceedings of advances in cryptology, eurocrypt’98
49. Yang Y, Zhang Y (2011) A generic scheme for secure data sharing in cloud. In: 40th
international conference on parallel processing workshops, pp 145–153, Sept 2011
50. Samanthula BK et al (2015) A secure data sharing and query processing framework via
federation of cloud computing. Inf Syst 48:196–212
Fine-Grained Access Control and Secured Data … 213
51. Boneh D, Waters B (2007) Conjunctive, subset, and range queries on encrypted data. In:
Proceedings of the 4th conference on theory of cryptography, TCC’07. Springer, Berlin,
pp 535–554
52. Hakan H, Iyer B, Li C, Mehrotra S (2002) Executing Sql over encrypted data in the
database-service provider model. In: Proceedings of the 2002 ACM sigmod international
conference on management of data, SIGMOD’02. ACM, pp 216–227
53. Hore B, Mehrotra S, Canim M, Kantarcioglu M (2012) Secure multidimensional range
queries over outsourced data. VLDB J 21(3):333–358
54. Wang S, Agrawal D, El Abbadi A (2011) A comprehensive framework for secure query
processing on relational data in the cloud. In: Proceedings of the 8th VLDB international
conference on secure data management, SDM’11. Springer, Berlin, pp 52–69
Author Biographies
Abstract Nowadays, social networking sites are very greatly used and are con-
tinuously growing at its peak. The extraordinary use of all the social networking
sites mainly Facebook, Twitter, LinkedIn, and Google Plus involve huge amount of
data transferred to public daily. This data transfer involves public information such
as personal information, education, professional, etc. which leads to security at
personal level. Let us see the comparative study of Facebook, Twitter, LinkedIn,
and Google Plus for security risk and how effective it is for well-being to society.
Keywords Social Facebook Twitter Gplus Linkedin Security
Public Private Government regulation Information Security
Optimization security measures Vulnerability
1 Introduction
T. Bhalodia (&)
Atmiya Institute of Technology & Science, Rajkot, India
e-mail: [email protected]
C. Kathad K. Zala
Ilaxo.Com, Rajkot, India
e-mail: [email protected]
K. Zala
e-mail: [email protected]
1.1 Facebook/GPlus
There are main three usual features for Facebook like capability to adding friends,
to change or modify status, and last one is implement application for execution
applications such as games and quizzes. A “Friend” means anyone on the Facebook
system whom you allow to see very different levels of personal and public infor-
mation, such as comments, birth date, jobs, photos, member of groups, and list of
other friends and relatives. An individual can play games online and update others
in day-to-day life.
Everyone can also notice friend’s friend, i.e., individuals, whom you have
officially became friend and may not met before, may have visualization keen on
everyone’s private situations and information.
There is update field which is at the pinnacle of the everyone’s Facebook, but the
main use of that field is that it allows the abuser to place anything similar to snippet
as regards several topics at any point. It has very parallel field, although it does not
agree to extra passage, and LinkedIn is not allowed for connecting associations/
images/videos with the keep posted. A little example of every user’s update is
posted by your any social networking site like Facebook friend. These all are
extremely classic:
• “Presently established a plane ticket proffer.”
• “Someone is tired of every one this cold winter.”
Even though that strength looks comparatively undamaging, the third position
could raise a little be anxious. Every user can tell all their friends and connected
links, i.e., all of your friends, which we do not be there at home used for a half year
[12]. This is like to attaching an indication on the main road and infusing
something.
Although the applications on social network may seem to be safe, along with in
actuality a good number probably it is safe and harmless, it is forever something
that can send harmful content to your computer/laptops. It is not right just to
Facebook, but there are same as like Facebook, additional social networking sites
which are associated with the Internet in universal situation, when you start
Comparative Study of Security Risk in Social … 217
1.2 Twitter
It is a live application like to Facebook and LinkedIn which allows you to post
comments which we say these days to tweet on some topics. Special users on the
network of Twitter can grow to be supporters of someone’s tweets related, like
everyone can receive the updates regarding the data or information which are sent
by them.
Study over the business ecosystems in Hungary by monitoring 6000 out of
20,000 Facebook users who publically displayed their employers. Then, they rep-
resented the complexity of connections graphically through a simulator. Also, they
transformed the overall graphical network into a relationship graph of employers. If
individuals are very related to each other in the network, then there is strong
bonding in relationship with each other. Making progress to the same framework,
Neunerdt et al. [3] proposed two algorithms for collecting and processing web
comments in context of social blogging. Agarwal [1] proposed his extraordinary
research work on “Prediction of Trends in Online Social Networks”. He expended
the “directed links of following” in the social media of Twitter to determine the flow
of information. This approach directed a user’s influence on others users that could
decide if the topic is stylish or viral in the social networking world.
1.3 LinkedIn
If user can utilize Facebook, LinkedIn, Gplus, Twitter, or else some other online
site for social networking, Internet banking or daily purchases, you must be
responsive of messages and emails which are argued to be since these sites but
actually the tricks may contain nasty content. I have received many emails that
claim to be from my personal bank, but they are actually sent by a spammer.
Spammers are there in the hopes of obtaining my users user id and password.
Claiming of emails of Twitter and Facebook invitations is now most common.
Emails and messages may still contain an attach RAR file or ZIP file that recipients
may use to unlock to observe which user is invited them and made a flow to open
the file. The attachment actually contains a worm; it may destroy the entire user’s
computer and user’s reputation on personal and organizational level.
218 T. Bhalodia et al.
2 Research Background
The paper refers to the definition of information security as given by IMS for
information security. It defines “securing the information from different threats in
order to ensure business confidentiality, reducing business risk, and increase ROI
and business opportunities” [4].
See Fig. 1.
It is very important to keep awareness of social media among the employees and
within the organization. There are many negative consequences for such exploring
to social media. Due to such study and effect of information leakage and by pre-
vious studies, we have concluded that employees and other individuals are very
Table 1 The percentage of threat in real world when exploring personal information to social
networking sites [6]
Social site Uses Risk percentage
(%)
Facebook Facebook allows posting of personal data 61
Twitter Twitter creates business relations and connections 17
LinkedIn Creates followers 4
Google Same as Facebook but with low risk due to its usability 40
Plus …
Myspace User’s space 18
There are biggest risk of security Facebook (61%), Twitter (17%), LinkedIn (4%), Google Plus
(40%), and Myspace (18%)
Comparative Study of Security Risk in Social … 219
well aware about the threat by leakage of organizational and personal information
to the outer world, as mentioned issues and percentage of sharing information might
lead to serious damages.
By considering the security problem caused by social networking sites, the
suggestions must be implemented by organizations and individuals. Such consid-
erations are more distinguished by SETA programs, with organizational policies
[2, 12].
Based on above discussions, it can be said that many users are aware of the use
of social media that is directly concerned with the security issues. However,
organizations and many awareness programs in the society play important role to
install this awareness to individual in order to protect them from leakage of valuable
information of personal and professional information, which may cause serious
security disaster.
References
6. Neunerdt M, Niermann M, Mathar R, Trevisan B (2013) Focused crawling for building web
comment corpora. In: The 10th Annual IEEE CCNC- Work-in-Progress, pp 761–765
7. Olsik J (2011) The ESG information security management maturity model. Enterprise
Strategy Group, Milford, Massachusetts
8. PricewaterhouseCoopers (2010) Security for social networking. pwc.com.au, Australia
9. Radianti J, Gonzalez JJ (2007) A preliminary model of the vulnerability black market. Society
10. Rowe FM, Ciravegna F (2010) Harnessing the social web: the science of identity
disambiguation. In: Web Science Conference
11. Sophos (2011) Security threat report: 2010. Sophos Group, Boston, Massachusetts
12. Star T (2012) Don’t become an ‘accidental’ outlaw. In the Star Online
A Key Based Spiral Approach for DNA
Cryptography
Abstract The present paper provides the conceptual framework on DNA cryp-
tography. A key based spiral technique is proposed which uses the concept of a key
to generate a spiral transposition to provide more data security than the existing
technique. The existing technique is not much robust against attacks and it uses a
fixed spiral. Various attacks can lead to data access to unauthorized users. The
proposed technique makes the transformation based on key. The performance
comparison between the existing, i.e., DNA sequence dictionary method for
securing data in DNA using a fixed spiral transposition and proposed technique,
i.e., a key based spiral approach for securing the data, shows that the proposed
technique is much better than the existing technique in terms of MSE, PSNR, and
percentage of total bit changed and provides more security than the existing
technique.
Keywords DNA cryptography DNA structure Spiral transposition
MSE PSNR Percentage of total bit changed A, T, G, and C
1 Introduction
Nowadays, security of information has become very important with the growth of
technical advancement and the use of internet spreading day by day at a very rapid
rate. New technologies are emerging in IT sector which give invitation to the
attackers and threats. Lots of things are affected by these threats and attacks like
bank account, social security, etc. Information should be known to the receiver
only, but because of the weak spots in the security system information gets vul-
nerable by exploiting the weakness of system as follows:
• The area where ciphers are stored,
• Strength of the used algorithm,
• Random number generator, etc.
Therefore, the main job of the designer who designs the security system is to
diminish the chances of threats that attack our system and exploit the weakness. The
modern cryptography algorithms like DES and MD5 are also in danger zone [1].
For this purpose, DNA computing provides helpful direction in solving different
kinds of problems in cryptography and security. In DNA computing, DNA is used
as an information shipper or carrier and modern technology of biology is used as a
fulfillment tool. DNA cryptographic system is very powerful against attackers.
DNA molecules are used for cryptography because of their massive parallels and
extraordinary or vast information inherent capacity or density [2–4]. DNA becomes
a perfect medium for data hiding because of its extraordinary storage capacity and
the capability of synthesizing its sequence in any desirable length.
The chances of threats and attacks are increasing at a very rapid rate because of the
rapid development in the technical advancement. So, the required security require-
ments are not satisfied with the traditional cryptography methods [5–7] DNA is a
natural carrier of the information, which stores the data in the form of nucleotides.
DNA can store any type of data whether it is image, text, audio, or video. The
information stored in DNA can be any email, password, banking details, organization
details, industry details, or any other personal or private information [8, 9, 4]. DNA
cryptography provides security of data inherited in DNA in the form of nucleotides. It
encodes the data or information in such an efficient way that it can be transmitted
through any open environment in a secure and efficient way. So, two major issues
present in the field of system security and cryptography are storage and security and
both of these issues get resolved by the DNA cryptography [5, 3, 10, 11].
2 Existing Technique
value is given to the each 8-bit binary data. Then, the corresponding DNA sequence
is given to these decimal values from the given dictionary table containing 64
values. But the main disadvantage of this approach is that it uses a fixed spiral;
therefore, there is no flexibility in this approach [4].
3 Proposed Technique
The proposed technique is based on the concept of key based spiral approach for
DNA cryptography where data as well as the key is taken as an input as shown
below in the flowchart of Fig. 2. The data and the key both are converted into the
binary form. After converting them into the binary form, compare the size of data
bit with the size of key. If the size of key is less than the size of input data, then the
key is repeated to get the size equal to the binary input data. Now, each bit of key is
checked out. If the bit in the key is zero, then the new encrypted data have the same
value as that of old data means no replacement is made. But if the key has 1 in the
bit position, then the new encrypted data will have a new element, i.e., total data
length- key length. Suppose there are 100 elements in the input data. If we are
checking the first bit of key and it is 0, then the new encrypted data have the same
bit input as of old data. Suppose, now we are checking the second bit of key and it
is 1, then @(100-1), i.e., 99th bit of the old data will go to the second position of the
new encrypted data, and so on. Now convert this binary data into the decimal value
and find the corresponding four-character-long DNA sequence into the dictionary.
0 1 0 1 0 1 0 0
0 1 0 0 1 0 0 0
0 1 0 1 0 1 0 1
0 1 0 1 0 0 1 0
0 1 0 1 0 0 1 1
0 1 0 0 0 1 0 0
0 1 0 0 0 0 0 1
0 1 0 1 1 0 0 1
0 1 0 0 0 0 0 1
0 1 0 0 0 0 1 0
0 1 0 0 0 0 1 1
0 1 0 0 0 1 0 0
0 1 0 0 0 0 0 1
0 1 0 0 0 0 1 0
0 1 0 0 0 0 1 1
0 1 0 0 0 1 0 0
0 1 0 0 0 0 0 1
0 1 0 0 0 0 1 0
0 1 0 0 0 0 1 1
0 1 0 0 0 1 0 0
Now, the binary representation of the key can be given in a matrix form
(Fig. 4).
(e) Step 5: If the key size is less than the plaintext size, then repeat the key binary
values to make the size of key equal to the size of plaintext. Therefore,
repeat the binary key (Fig. 5).
(f) Step 6: Now make a spiral based upon the logic that if there is 0 at any bit
position in the key then the new encrypted data has the same value as
that of old data. But, if there is 1 in the key bit position, the data bit is
replaced by a new data bit. Suppose there is 1 on the first position in the
key and total data length is 64, then we will put (64-1) 63 element of the
A Key Based Spiral Approach for DNA Cryptography 227
0 0 0 1 0 1 0 1
0 0 0 0 1 0 0 0
0 1 0 1 0 1 0 1
0 0 0 1 0 1 1 0
0 0 0 1 0 0 1 1
0 1 0 0 0 1 0 0
0 0 0 0 0 0 0 0
0 1 0 1 1 1 0 1
old data into the first position of the new encrypted data. Therefore, a
spiral based on this logic can be represented as (Fig. 6):
(g) Step 7: Now, the spiral’s equivalent ASCII values and DNA sequence using the
DNA sequence dictionary are shown in (Table 3):
Therefore, from the above illustration, the encrypted text for the plaintext
“THURSDAY” using a key “ABCD” is “ATTAAATCTTTAATTTATAG
TAACCCCCTTCA”.
4 Performance Comparison
The performance comparison has been made between the existing and the proposed
technique on the basis of PSNR, MSE, and percentage of total bit changed. It is
clear from the results that the proposed technique is much better than the existing
technique in terms of PSNR, MSE, and total percentage of bit changed. The pro-
posed technique provides the low MSE and low total bit changed and high PSNR
than the existing method.
228 Ekta and A. Singh
PSNR (Peak Signal-to-Noise Ratio): It is the measure of quality of picture with the
comparison of cover image and stego image. If PSNR is higher, it means the
performance of technique is also higher [12, 13].
MSE (Mean Square Error): It is defined as the square of error between the stego
image and the cover image. It should be low for better results [14, 11].
Percentage of total bit changed: The total percentage of the bit change represents
the amount of bits change in the cover data to get the stego data. The lower the bit
change, the better is the technique [12, 13].
Various graphs generated using Matlab® and program codes are shown in
(Figs. 7, 8, and 9).
Fig. 7 MSE versus message length for proposed and existing technique
8
(
PSNR dB
PSNR by
6
Present
)
4 Method
2 PSNR by
Proposed
0 Method
3 4 5 6 7 8 9 10
Message Length
Fig. 8 PSNRS versus message length for proposed and existing technique
A Key Based Spiral Approach for DNA Cryptography 229
40
Changed
30
20 % of total bit changed by Present Method
10
% of total bit changed by Proposed Method
0
3 4 5 6 7 8 9 10
Message Length
Fig. 9 Percentage of bit changed versus message length for proposed and existing technique
5 Conclusion
The existing technique is having a number of limitations such as use of fixed spiral,
high MSE, low PSNR, and high percentage of total bit changed, due to which it
doesn’t fulfill the security requirements. Keeping in view to avoid such types of
problems, the new technique is proposed which makes use of concept based on
key based spiral transposition where the encryption is based upon key; different
keys can be used to encrypt any type of data. Therefore, security is enhanced due to
key flexibility system. The logic used in the key is based upon the existence of 0 or
1. If there is 1 in the key bit position, the data bit is replaced by a new data bit.
Suppose there is 1 on the first position in the key and total data length is 100, then
we will put (100-1), i.e., 99 elements of the old data into the first position of the new
encrypted data. Therefore, the logic used here is better as compared to the earlier
approach because it is based upon a key flexibility system, i.e., it can use a key or
different keys to encrypt the data. The three main parameters, i.e., MSE, PSNR, and
percentage of total bit changed also give us better results. The value of MSE is low;
the lower the value of MSE, the better the result. The value produced by the PSNR
is also high; the higher the value of PSNR, the better the results. The total number
of bits changed also gets reduced so that the attacker cannot differentiate between
the actual data, i.e., DNA and the encrypted data. Therefore, in comparison to the
existing technique, this proposed technique has a good flexible key based
encryption method and proposed method also produces lower MSE, higher PSNR,
and total percentage of bits changed get reduced than the present method. Hence,
the proposed technique is better than the existing technique. One other benefit of the
proposed method is also the flexibility of key used.
References
1. Jacob G, Murugan A (2013) DNA based cryptography: an overview and analysis. Int J Emerg
Sci 3(1):36–42
2. Chen J (2003) A DNA-based, bimolecular cryptography design. In: ISCAS’03 proceedings
3. Gehani A, LaBean T, Reif J (2004) DNA-based cryptography. Lecture Notes in Computer
Science, Springer
230 Ekta and A. Singh
4. Jain S, Dr. Bhatnagar V (2014) A novel sequence dictionary method for securing data in DNA
using spiral method and framework of DNA cryptography. ICAETR
5. Soni R, Prajapat G (2013) A modern review on DNA cryptography techniques. Int J Adv Res
Comput Sci Softw Eng 3(7)
6. Amin ST, Saeb M, El-Gindi S (2006) A DNA-based implementation of YAEA encryption
algorithm. In: IASTED International Conference on Computational Intelligence
7. Pruthi Y, Dixit S (2014) A comparative study on DNA cryptography. IJARCEE 4(5)
8. Kahate A Cryptography and network security (3rd edn). McGraw-Hill
9. Shyam VMM, Kiran N (2000) A novel encryption scheme based on DNA computing. In:
14th IEEE International Conference, Tia, India
10. Borda ME, Tornea O (2010) DNA secret writing techniques. In: IEEE conferences
11. Kumar S, Chakraborty S (2011) Image steganography using DNA sequence. AJCSIT 1:2
12. Khalifa A, Atito A (2012) High-capacity DNA-based steganography. In: The 8th International
Conference and informatics and Systems, IEEE
13. Torkaman MRN, Kazazi NS, Rouddini A (2012) Innovative approach to improve hybrid
cryptography by using DNA steganography. IJNCAA 2(1)
14. Tulpan D, Regoni C, Durand G, Bellivean L, Leger S (2013) A hybrid stegano-cryptographic
approach for data encryption using randomized error correcting DNA codes. HINDAWI 2013
Author Biographies
Ekta is currently pursuing PhD (2018) in Computer Science and Engineering from Bhagat Phool
Singh Mahila Vishwavidyalaya, Haryana, India, a governement university.
Prof. Ajit Singh is working as Professor in Dept. of Computer Science and Engineering of Bhagat
Phool Singh Mahila Vishwavidyalaya, Haryana, India, a governement university.
Permission-Set Based Detection
and Analysis of Android Malware
Abstract Smartphone industry has become one of the fastest growing techno-
logical areas in the past few years. The monotonic growth of Android share market
and the diversity among various app sources besides official Google Play Store has
attracted attention of malware attacker. To tackle with the problem of increasing
number of malicious Android app available at various sources, this paper proposes
a novel approach which is based on feature similarity of Android apps. This
approach has been implemented by performing static analysis to extract the features
from an APK file. Extracted features are useful and meaningful to make efficient
training system. This paper proposes a permission-based model which makes use of
self-organizing map algorithm. The implemented approach has been analyzed using
1200 heterogeneous Android apps. The proposed approach shows improved results
for TPR, FPR, and accuracy.
1 Introduction
The shares of smartphone market are increasing day by day. Numbers of new
smartphone companies are coming in market with new models having number of
new and extra features. Although smartphone industry is a tough industry, if one
wants to become a part of this industry, they have to come with little bit more that is
just good. This can be provided by enhancing smartphone processor, operating
system, battery life, storage capacity, screen resolution, and many more. In addition
to this, smartphone has inbuilt apps which provide extraordinary features.
Simultaneous execution of many app creates a burden on operating system for
2 Related Work
The number of concept and technique was proposed to reduce the growing amount
of Android malware. To gain knowledge about malware propagation, a detailed
study of related work is needed. A number of survey papers were written to pay
attention on malware detection, analysis, and propagation with their cause and
effect that was provided in studies of [3–6]. The Cooper et al. [3] Android apps was
developing quickly for almost covering gaming, entertainment, adventure, educa-
tion, social media, businesses, lifestyle, and other day-to-day activities. With quick
development of these apps, developer had not provided more attention to security of
these apps. Hence, a solid understanding of malware characteristics was needed.
This helped to prevent many unwanted consequences present in the app. [4] told
that malware was not only related to abnormal execution of a program. There were
big incentives for writing malware [5]. Basic reason of all malware was to harm the
users by any means. But all those malware were not shared similar characteristics.
Based on malware characteristics, Android malware falls in 49 different malware
families [6]. They gave first mobile malware survey and analyzed a total of 46
samples of iOS, Symbian, and Android with their incentives. Other Android-based
analysis and detection approaches [7–9] were existed in literature. The authors
utilized only permission set to detect and analyze the Android malware [7]. They
developed a two-layered permissions-based detector. This detector utilized only
Permission-Set Based Detection and Analysis … 233
requested features and used-features of Android apps [8]. They gave PUMA to
detect Android malware also based on permission set [9]. They studied relationship
of permissions requested by Android apps. This study was intended for
non-malicious Android apps only.
To detect malicious apps, the proposed approach requires low-level feature visu-
alization with high-level feature similarity that helps to determine typical indica-
tions of malicious activity. Many researchers implemented their approaches by
knowingly class label in advance. To implement the proposed approach, we do not
need any class label in advance, as we employed unsupervised clustering technique.
The dataset has been collected from official Google Play Store [10] and android-
sandbox.net [11]. The implementation process is as follows:
(a) Android app reverse engineering: In the first step, a Java-based APK tool has
been used for automated reverse engineering of APK files. This tool provides
readable Android Manifest.xml file, multiple Smali files, and other resource
subfolders (Sect. 3.1).
(b) Broad static analysis: In the first step, given Android application is inspected
and extracts different features from the Android manifest and Smali files
(Sect. 3.2).
(c) Embedding in vector space: The extracted features have been mapped into a
vector space separately for each model, where patterns and combinations of the
features can be analyzed on the similarity bases (Sect. 3.3).
(d) Self-organizing feature map based visualization: Self-organizing feature map
make cluster based on their feature similarity. The embedding of the features
enables us to identify malware using SOM, which provides better visualization
of features (Sect. 3.4).
The collected Android apps are in .APK format which are not directly readable to
the users. To make these apps readable, we have to transform these apps into
readable format. This transformation of APK files has been achieved by doing
reverse engineering. APK tool [12] is used for transformation of APK files. APK
tool transforms app nearly to its original form. One can modify the app and rebuild
that app. It debugs Smali code step by step [15]. Generally, modifications are made
to add some features. But attacker adds malicious code and rebuilds the app to harm
the user.
234 A. Sharma and A. Doegar
All smartphones which are based on Android operating system have different
capabilities, for example, some devices support cellular data networks while others
only support Wi-Fi. Android deals with variety of features. The static analysis
inspects given Android app without execution and extract as many as possible
permissions. Every Android app must include a manifest file called as
AndroidManifest.xml. This single file provides complete information that supports
the installation and later execution of the application. Android app needs some
functionality like reading SMS, sending SMS, making call, using camera, etc. It is
mandatory to protect this functionality from unauthorized use. Android permissions
have been created to protect these functionalities. Every app has a set of permission
which user accepted at install time. The extracted permission has been helpful and
meaningful to construct the model. The appropriateness of extracted features affects
the result. The features have been extracted from Android manifest file of an
app. For an example, if Android app wants to access the complete information of a
network, the following is declared in Android manifest file:
<uses-permission android: name = “android.permission.ACCESS_NETWORK_
STATE”/>
To declare a permission <uses-permission> tag is used.
The extracted features have been mapped into a vector space. An input vector has
been made corresponding to an app permission set. The requested permission has
been expressed in the form of a bit string. Every app permission has been stored in
binary (0, 1) format. If feature vector is represented by F, then
1 if and only if the ith permission requested
Fi ¼
0 otherwise
The permission-based model has been trained using Self-Organizing Maps (SOM).
SOM is trained using unsupervised learning to produce a low-dimensional (typi-
cally two-dimensional) view of data [13].
SOM Algorithm [9]
Step 1: Initialize neuron weights wi ¼ wi1 ; wi2 ; . . .; wij 2 R. To initialize the
Neuron weights, random numbers have been used.
Step 2: Load an input pattern x ¼ x1 ; x2 ; . . .; xj 2 M. Here, M can be any three
models M 2 [Permission-based]. Distance between pattern x, and each neuron
weight wi, has been calculated and winning neuron or best matching neuron c has
been identified as follows:
kx wc k ¼ minfkx wi kg
where
where
rc and ri are positions of neurons and on the SOM grid
a(t) learning rate function
(t) defines the width of the kernel
Both a(t) and r(t) decrease monotonically with time.
236 A. Sharma and A. Doegar
nben!ben þ nmal!mal
ACC ¼ ð3Þ
nben!ben þ nben!mal þ nmal!ben þ nmal!mal
4 Performance Evaluation
80%
Percentage
60%
TPR
40%
FPR
20% ACC
0%
TPR FPR ACC
Performance Measure
Figure 1 depicts the results that show the average detection rate for new Android
malware by Android malware. The permission-based model achieved good detec-
tion rate. In permission-based model, filtering is done only on the bases of per-
mission set which gives TPR, FPR, and ACC 91, 9, and 90%, respectively.
5 Conclusion
This paper proposes a novel approach for detecting potential Android malicious
apps using Self-organizing Maps (SOM). Self-organizing map is an unsupervised
clustering method; no prior information about class labels has been required. SOM
detect and analyze new Android apps based on their feature similarity. The
permission-based model has been investigated to detect malwares present in
Android apps. In order to build the permission-based model, all the permissions
have been extracted from a large number of benign and malicious APK files. APK
files are not human readable file and feature cannot be directly extracted from these
APK files. The APK tool has been used to transform the APK files into readable
format and to extract features. The permission-based model has been constructed by
extracting features from 400 benign apps and 400 malicious apps. To test
permission-based model, total 400 apps (200 from benign dataset and 200 from
malware dataset) have been used.
The evaluation of models is done using performance measures such as True
Positive Rate (TPR), the False Positive Rate (FPR), and Accuracy (ACC).
The TPR, FPR, and ACC of model are 91, 9, and 90%, respectively, which is better
than existing approaches.
238 A. Sharma and A. Doegar
References
Author Biographies
Abstract Geographic Information System (GIS) plays a vital role in many appli-
cations especially in military operations as they need to be spatial in nature.
Successful application of military operations demands for accuracy of information
and quick decisions taking steps. GIS has now become the most powerful medium
for sharing of military information to officers and commanders. In the era of digital
communication, officers use GIS to deliver their strategic plans to intended officers
[5]. GIS has proven to be an excellent tool for enforcement and deployment of
security mechanisms in military applications and to deliver confidential information
at distant locations. In our proposed system, we will introduce a new mechanism to
protect GIS data carrying confidential and sensitive data for military and army
purpose by combining two of the cryptography algorithms: Advanced Encryption
Standard (AES) and RSA with digital watermarking techniques.
1 Introduction
Experts have long been recognized the importance of GIS in military and com-
mercial application. The GIS data has two important properties. First, the effort it
takes to put it in a suitable form for use in the GIS applications. This effort increases
its cost. Second, GIS data contains confidential and sensitive information most of
the time and it needs to be kept away from unauthorized users. Two possible threats
for GIS data are as follows:
2 Organization of Paper
Paper is organized into seven sections. Section 3 gives applications and limitations
of cryptography and digital watermarking. In Sect. 4, we discussed security issues
and threats while using GIS data with respect to the past work. We introduced and
explained our proposed system with the help of flowchart in Sect. 5. At last, the
paper is concluded by its future work.
Cryptography, digital watermarking, and many other technologies have been used
to handle security threats. Each of these technologies has its limitations and has
been long used as a weapon to solve security and authentication problems related to
data transmitted over network. Both of these have diversified applications and
usage with different objectives.
Cryptography tries to take care of three important properties of information
including confidentiality, authenticity, and integrity, while it is being transmitted
over public network. It is the method of encryption of original data at sender side
using key and algorithm before being transmitted over Internet and do reverse of the
same process at receiver side. Encryption is the process of converting a readable or
meaningful data in an unreadable and meaningless form. Many algorithms like
AES, RSA, Hashing, etc. used for the same purpose. It is also used for the purpose
of sharing secured data over unsecured network. The efficiency of cryptography
depends on key management and its distribution and not on the algorithm used and
this is one of the biggest security threats with this technique.
Digital watermarking is the method of hiding a digital information into digital
signal like an image, audio, or video signal itself. One of the mostly used appli-
cations of digital watermarking is owner identification. To identify the owner of
specific image or song, copyright information is embedded in the image or song
Three-Level GIS Data Security: Conjointly … 243
4 Related Work
From the past many years, GIS has been used by government agencies to transfer
information. Earlier, professionals used to identify threats, plan resource deploy-
ments, and map potential action and contingency plans with the help of GIS. Also,
for drawing and printing maps and for building of information desktop, applications
were widely used. However, nowadays, the GIS platform allow users the ability to
access confidential information and to use of maps in any easy manner in 24 7
from anywhere to anywhere and also on any network [3]. Being of its capability to
deliver confidential and secret information, GIS is used by military forces in a
variety of applications including terrorist activities monitoring, remote sensing,
borderlines monitoring, order enforcement at battlefield, etc. The work that has
been done in this area is mostly based on digital watermarking [6, 7]. Watermarking
is a process of obtaining a digital watermarked file by embedding hidden infor-
mation (watermarking pattern or watermark for simplicity) like copyright string in a
dataset without producing perceptible changes in the data using a suitable water-
marking algorithm [1].
5 Proposed System
image is extracted and used [2]. The sequence of steps taken at sender and receiver
side is shown as (Figs. 1, 2).
5.1 Explanation
The whole process of encryption and decryption passes through three stages of
encryption at sender side and decryption at receiver side. At first stage, message
which is to be sent by embedding it into an original image is encrypted using
128-bit AES algorithm in parallel extraction of least significant bits from original
Three-Level GIS Data Security: Conjointly … 245
5.2 Explanation
Similarly, at receiver end decryption process will be applied. First, the complete
encrypted product, p, gets decrypted using RSA and then decompose into two parts
x and En(i) (selected bits of the coordinates). Now, x will be decomposed by
applying decomposition function on x, and hence we will get Ex(f), En(w), and
copy of original image, j. Next, both encrypted message, En(w), and encrypted
image, En(i), get decrypted using 128-bit AES algorithm. This decrypted copy is an
another copy of the same original image, say i. Now by comparing the extracted
feature (bits) with features of both of these copies of original image, i and j, we can
easily judge whether the image being received is distorted or not.
Three-Level GIS Data Security: Conjointly … 247
6 Conclusion
Under this research paper, we proposed a system that provides security at three
levels. If an intruder is somehow able to decrypt the data at any of these levels, then
it will be very difficult to decrypt at all levels. We have provided the security mix of
symmetric and asymmetric cryptography that increases the security of the system.
The proposed system can be used in the applications where sensitive information is
needed to be transferred.
7 Future Scope
In this research paper, we have just proposed a system that could be implemented
for providing efficient security to the organizations, businesses, military, medical,
etc. Although the system provides security at three stages using symmetric as well
as asymmetric algorithm, however, the use of other asymmetric algorithm for final
stage could make system more secure, efficient, and accountable. Since the pro-
posed system does not support any experimental data set, hence, it needs to be
implemented for its actual result with strong mathematical foundation, comparing it
with other algorithms on the basis of parameters like performance, efficiency, and
complexity.
References
Avijit Dutta
Abstract The subject security has wide coverage and it is growing with every
passing day. As civilization progressed from Agrarian to semi-industrialization,
advanced industrialization and finally to present ICT (Information and
Communication Technology) age, concerns for security are increasingly taking in
all objects from physical to digital. It augmented apprehensions from losing
material wealth to most abstract entities like wealth of knowledge in digital form.
Today’s technology allows wired and wireless access to tangible and intangible
resource-built ups (material to digital), digitally, and steal the same if need arises.
The riddle is to defend our own resources from the rapacious hand of ubiquitous
computing and communicating technology evolved by us. The art and science of
hiding and securing precious resources from possible predators in physical or
digital forms make it complex and challenging. The enigma remains in the fact that
predator uses same technology and at times also makes rule that prevails over
others.
1 Introduction
Technology integration and its standardization have put civilization on fast track.
From ‘agrarian’ to ‘semi-industrialization’, ‘industrialization’, ‘advanced industri-
alization’ and finally to ‘digital age’, the journey so far has been exciting.
Innovations across different subject areas cooperate amongst themselves to make
ways for new novelty. Weiser [1, 2] may have closely followed advances in
computing hardware, system software and programming techniques during 90s
to visualize the phenomena of ubiquitous computing, which now is a reality.
A. Dutta (&)
NIC, New Delhi, India
e-mail: [email protected]
Broadly, three factors have driven computing technology to ubiquity. First to name
is ‘Miniaturization’, which is a trend to manufacture ever smaller mechanical, optical
and electronic products and devices. Second to mention is ‘Standardization’, which is
the process of developing and implementing technical standards that helps to max-
imize compatibility, interoperability, safety, repeatability or quality. Third to mention
is ‘Digital Communication’, which evolved over packet switch networking tech-
nologies, mostly adhering to TCP/IP protocol standards. This allows data exchange
between computing devices over wired or wireless network. At the advent of TCP/
IP-related protocol like HTTP (Hypertext Transfer Protocol), World Wide Web
(WWW) became a reality leading to web 1.0 paradigm, which allowed viewing vast
amount of static information on web, advancing data disseminations practices,
leading to dot-com era. Initial enthusiasm died down as viewers could not participate
in the process, thus followed occurrence of dot-com burst. Web 2.0, which is
interactive, revived web and took it to today’s state of booming activities where
everyone is keen to participate. In exponentially expanding web scenario, the
exemplar that may follow web 2.0 is a subject of any one’s guess now! To some, it is
web 3.0, simply as next version standard, with more advanced technical facilities. For
others, it is ‘Web Square’, the name and concept popularized by Tim O’Reilly and
John Battelle. Progression of events allowed Tim O’Reilly, at a later date, to talk
about IoT (Internet of Things) and collective intelligence [3]. He, during early years
of twenty-first century, could visualize flooding of Internet usages with sensors and
devices leading all to an era of nomadic and yet interactive WEB [1, 2, 4–6].
It was expected that the number of such devices would grow exponentially to
guide technology to next-generation usages. These sensors and devices singularly
termed as IoT are designed to add intelligence to everything from commonplace
consumer items, home appliances, private or public utility systems, industrial items,
healthcare system, education, agriculture and everything in between, even to rail-
road ties on big or small deals. ‘IoTs’ collects and broadcasts data across networks,
enabling the data to be analysed on it or remote servers to add values and share.
This approach changed the very way life and business processes were hitherto
accomplished, leading to an archetype shift from physical to digital course of
functioning [7–11].
Technology advances ushered era of first, second, third and fourth generations of
computing. During this period, human–computer interactions shifted from ‘One
Machine many users’ to ‘One user One Machine’ and finally to ‘Many Machine
Many User’ setups. Digital computing stepped out from closed realm of scientists
and academicians to arrive at the doorstep of common users. As discussed earlier
technology integration, its standardization and digital communication steered us to
the era of WWW and Internet. Broadly, evolutionary path of Internet can be viewed
as follows—from years 1969 to 1995 it belonged to hardcore technocrats and
scientists, from 1995 to 2000 it belonged to geeks, from year 2000 to 2007 it
became Internet of masses, from 2007 to 2011 it turned as Internet of mobiles and
from 2012 and days beyond it may evolve into the era of IoT. It may be opined that
emergence of web 1.0 (static web) occurred during Internet of geeks and web 2.0
Digital Security: An Enigma 251
2 Collective Intelligence
The concept of data and the process of its collection, collation and dissemination
have changed largely in the era of web 2.0 [7, 8]. Today, apart from texts, digits,
audio and video, photographs too are taken to mean as data. Keyboard now is not
the only means for data incorporation, interpretation and interaction with digital
objects and Internet. Omnipresent smart devices can look, feel, sense, photograph
objects and store them within a split second instruction at any desired location,
really smartly [3, 6].
Technology miniaturization, standardization and large-scale product manufac-
ture are bringing down the cost of computing and communicating. This has helped
a wide range of computing and communicating devices in terms of size and per-
formance like servers, desktop, laptop, palmtop, smartphones, wearable devices,
etc. to be available in the market. These devices are also armed with seeing,
listening, recording and storing capabilities, which cater to extensive range of data
processing and disbursing needs, helping to bring most on board. These devices
with an identity can be linked amongst themselves and numerous other small or
large smart digital devices, termed singularly as IoT, as discussed earlier, over
varied choices of connectivity options like broadband, Wi-Fi, R/F, Bluetooth, etc.
[14, 15].
The depiction in Fig. 1 (IoT Scenario) attempts to present a window view of the
situation arising out of the increasing presence of IoTs. This helped to enhance the
mass base of smart devices usages. Digital devices are capable now to communicate
intelligently amongst themselves and others in forms like M2M (Machine to
Machine), M2I (Machine to Infrastructure) and M2E (Machine to Environment) in
real time, process data at nodes or cloud deciding almost autonomously and present
the most up-to-date information to us so that we can make the best decisions.
252 A. Dutta
Benson Tao observes that present efforts towards building smart, connected,
autonomous and contextually aware devices around the IoTs will prove to be catalyst
for a change, leading to general betterment. As it turns out, IoT is a very broad
concept, which includes all kind of wearable, carriable, attachable and implantable
and everything in between devices that associates with us in our daily coir.
Interestingly, O’Reilly [3] envisaged today’s Internet as a new born kid, who
looks, touches and feels about the things around with the help of various sensors
(being carried by us), like mobile phones and smart devices, to gather data in audio,
video and text form and processes them to attain a higher state of awareness. It is
increasingly getting intelligent with information gathered by sensors ubiquitously
strewn around, in both static and mobile state and maturing incrementally like any
living objects, though as a virtual entity. In return, it shares the collected data,
information and knowledge whenever these are asked for, inform of an organized
query, over digital network, establishing the concept of collective intelligence.
Worldwide efforts are on to bring most on board, to enrich the process of collective
intelligence and get maximum benefit out of it. Well, there is dark side of this
process too, which is being discussed in following segments.
3 Emerging Challenges
In keeping with Mark Weiser’s view of ‘ubiquitous computing’ concepts, one may
find that Computing and Communicating (C&C) emerged as profound technology
in this era, which has associated with our day-to-day life processes inseparably and
continuing expansion process of its presence exponentially with smart devices
termed as IoTs [1–3]. These phenomena are making fast inroads in our daily
activities. Broadband routers offer Internet access to devices through Wi-Fi and
Ethernet connections to make today’s home network. Appliances like laptops,
desktop computers and mobile devices, such as phones and tablets can get onto
Internet through broadband router. With the IoTs finding their ways into the homes,
innumerable new devices are produced that can connect to the same network. These
devices are of two types; the first ones get connected through formal networking
technologies as discussed earlier. Others may use different wireless technologies
that suite device needs, conforming to lower energy consumption or ad hoc network
coverage protocols. Nevertheless, everything is connected to the local network and
can communicate freely with one another. Connections to the Internet are directed
through a central router, which may (or may not) always contain basic firewall
filtering functionality [9–11].
It may be known that connected version of different devices, participating in
day-to-day activities, gets onto same network without essential security consider-
ation. Despite increasing acceptance of IoTs, no standards have been planned so far
for the use of these innumerable devices and sensors. They are almost on their own
in the process of establishing connection, exchanging and processing information
on instruction from numerous lawful or unlawful owners. Along with many goodies
that computing ubiquity presents, the offered challenges lie in the fact that the ‘IoT’
today is an abstract collection of uses and products without common agreement or
disagreement on mode of functioning. So, everyone does it their own way, often
poorly, compromising security of connected devices as it greatly lacks an estab-
lished concept of implementation and use. A study of security major like Symantec
Corporation seems to have found that currently there is no single standard protocol
in IoT and ‘security’ is not a word that gets strongly associated with this category of
devices, leaving its consumers potentially exposed [9–11]. The ‘enigma one’ lies in
the fact that these challenges are our own creation and we are forced to face them.
As information highway is being accessed by one and all, gradually concerns are
gaining ground about the co-travellers with whom this highway is being shared!
Symantec, after analysing 50 home devices, during year 2014, has observed that
none of the devices used strong password, enforced mutual authentication practices
or applied defence mechanism against brute-force attacks [10, 11]. It also has found
mobile apps generally do not use Secure Sockets Layer (SSL) to encrypt com-
munications to the cloud. The family of IoT devices possesses vulnerabilities,
which are much in common. Potential weaknesses in authentication and traffic
encryption could badly affect IoT systems. These facts though well known to the
security industry, mitigation processes are not taken good care off.
It is generally felt that IoT vendors need to do more on security before marketing
their product universally, leaving millions of people at risk of cyberattacks. This
leads a feeling that ‘IoT security is still a pipe dream’ [9–11, 15]. The ‘Enigma
Second’ lies in the fact that IoTs are being produced in large numbers with com-
prehensive knowledge about associated security hazards.
The digital security challenge mitigation begins with stopping innumerable
entities approaching digital resources over data communication network, to verify
their credentials and allowing passage, if found acceptable, denying it otherwise.
254 A. Dutta
The process gets multifaceted as advancing objects grows in number and form,
which taxes time and computing resources of approached entity. More often than
not objects seeking access to resources are large, interactive and at times deceptive.
Objects approaching resources constantly change form and advancing tactics to
match the defence mechanism with the intent to crack the same. Real-time detection
of specifics on attack vector is difficult and this leads to security breach.
Authentication and authorization become important at such instances. Digital
resources can be protected with cryptographic techniques and establishment of PKI
(Public Key Cryptographic Infrastructure) system. Steganographic techniques
allowing enveloped exchange of document also come handy for secured data
exchange. Cryptography and steganography putting together can provide robust
defence mechanism against predators during digital document exchange. Enigma
Third’ lies in the fact that for the defence of our ever evolving resources, eternally
new mechanism needs to be explored. In the following section, collective effect of
enigmas around ‘Digital Security’ is further discussed.
4 Enigma
Collectively, the digital security enigma lies in the fact that we are defensive against
our own creation and in a way we are creating our own space for both security and
insecurity. Adding to woes are the facts that lessons on computer hacking are
included officially in course curriculum of many national and international uni-
versities. Today, the attacker and defender use same or equivalent technologies and
at times they appear like either side of the same coin. Like for cryptography there is
cryptanalysis, for steganography there is steganalysis and so on. Moving on one
may even find that the perception of cybercrime is relative to geographical or
political jurisdiction. The inherent view that hacking others network is fair, getting
hacked is not, is scaring [8].
Financial sites of many institutions and well-offs are recurrently hacked by less
fortunate for instant monetary gains using advanced C&C technologies. Scientific
and Defence research sites of many advanced countries are being routinely intruded
these days for a fast track course to new knowledge, while gainers appreciate the
act, losers strongly denounce it. This has compelled many original equipment
manufacturing countries to embed cyber sniffing tools, in both hardware and
software systems, which are difficult to shake off, so as to pre-empt movement of
cyber predators [9–11]. It ensures (!) security breach even with best defence
mechanism up front as the attack can be initiated from either side of the system.
Fact remains that resourceful and militarily powerful countries cyber-snoops
friends and foes including close allies, all alike. IoTs have made the process even
simpler. These devices have made even our residential places vulnerable. Gartner
research predicts that there will be more than 2.9 billion connected IoT devices in
consumer smart home environments in 2015. These connected devices could pro-
vide a much larger surface for attackers to target home networks. IoTs are wearable,
Digital Security: An Enigma 255
5 Analysis
The journey over Internet for knowledge and wisdom at this moment is open to all,
which is expected to lead humanity to freedom from dogma, biases,
short-sightedness, etc., the factors that slow down the process to become a superior
entity. Plethora of web applications and mobile apps are being developed to ease
the use of Internet; wherein, required technical knowledge of computing and
communicating are minimal. Of late it is being observed that this freedom is being
256 A. Dutta
Table 3 Network readiness and cyber security status of top 10 countries in the world
S. Country GITR Network ITU Cybersecurity ITU Cybersecurity
No. Name Readiness Preparedness Preparedness Index on 10
Index (A) Index (B) Point Scale (B10)
1 Finland 6.04 0.618 6.18
2 Singapore 5.97 0.676 6.76
3 Sweden 5.93 0.647 6.47
4 Netherland 5.79 0.676 6.76
5 Norway 5.70 0.735 7.35
6 Switzerland 5.62 0.353 3.53
7 United 5.61 0.824 8.24
States
8 Hong Kong 5.60 0.618 6.18
SAR
9 United 5.54 0.706 7.06
Kingdom
10 Korea Rep 5.54 0.706 7.06
Source [16, 17]
6 Conclusion
With a limited scope, deliberations so far have indicated that there is high corre-
lation between ‘Network Readiness Index’ and ‘Cybersecurity Preparedness Index’.
Though absolute security is not achievable in today’s scenario, mainly because of
the fact that same technology and related standards are being used by both attackers
and defenders, remaining oblivion to security issue may be catastrophic. Digital
networks have been opening up precious resources to one and all at the backdrop of
the debate on ‘Net Neutrality’; thus, combinations of security options, with focus on
digital network, may help in making a strong security module to enhance safety.
7 Future Scope
It has just been conceived that ‘Digital Security’ aspect increasingly getting
probabilistic and security model needs to be evolved to control ‘Enigma’ with an
aim to establish enhanced safety. In this context, data from more countries needs to
be included to make the study further accurate. Apart from factors like ‘Network
Readiness’, associations of other factors like country-wise Knowledge Index (KI),
knowledge economy index, ICT index, etc. with cybersecurity preparedness index,
may be explored individually and collectively to evolve a reliable security/safety
model that assures safe network usages.
258 A. Dutta
Acknowledgements Contributions of theorists who can look into the future and guide science
and technologies beyond horizon are deeply acknowledged.
References
1. Weiser M (1991) The computer for the 21st century. Scientific American, Sept 1991,
pp 94–104
2. Weiser M, Brown JS (1996) The coming age of calm technology. Xerox PARC, 5 Oct 1996
3. O’Reilly T, Battelle J. Web squared: web 2.0 five years on; special report
4. Kleinrock L. Nomadic computing—an opportunity CCR 4/95
5. Burgin M, Eberbach E (2012) Evolutionary computation and the processes of life. ACM
Publication
6. La Porta TF, Sabnani KK, Gitlin RD. Challenges for nomadic computing: mobility
management and wireless communications. Bell Laboratories
7. Avijit D. Knowledge ubiquity in web 2.0 paradigm. Innovation in information system and
technology. ITCDC ‘09 Macmillan Publications, pp 234–238
8. Avijit D. Digital security: a moving target. Int J Electr Electron Comput Sci Eng. Special issue
—TeLMISR 2015. ISSN: 2348-2273
9. Barcena MB, Wueest C (2015) Insecurity in the Internet of Things. Symantec, security
response, version 1.0, 12 March 2015
10. Symantec, ISTR, April 2015, vol 20
11. Symantec, Insecurity in Internet of Things, version 1.0, 12 March 2015
12. Cortada JW, Marc GAMLN How nations thrive in the information age. IBM Institute for
Business Value, IBM Global Business Services
13. Kephart JO, Chess DM (2003) Autonomic computing. IBM Thomas J. Watson Research
Center, IEEE Computer Society
14. https://bensontao.wordpress.com/2013/10/06/vivante-internet-of-things
15. http://securityaffairs.co/wordpress/34974/cyber-crime/iot-security-symantec.htm
16. INSEAD (2014) Global information technology report
17. ITU (2015) Global cyber security index & cyberwellness profile report
18. Lytinen K, Yoo Y. The next wave of nomadic computing: a research agenda for information
systems research. Working papers on information systems, Sprouts. ISSN: 1535-6078
19. Cousins KC, Robey D. Human agency in a wireless world: patterns of technology use in
nomadic computing environments. Information and Organization; Science Direct
20. Venkatasubramanian K, Gupta SKS (2006) Security solutions for pervasive healthcare. P1:
Binaya Dash, 8 Dec 2006, vol 11:58, pp AU7921–AU7921˙C015
21. Kleinrock L. Nomadic computing. Computer Science Department, Los Angeles
22. Davis RM. Evolution of computers and computing. Science 195
23. Satyanarayanan M. Pervasive computing: vision and challenges. School of Computer Science,
Carnegie Mellon University
24. TechTarget, Security Media Group. Information security, October 2014, vol 16, no 8
25. http://www.slideshare.net/MhaeLyn/iot-30545508
Digital Security: An Enigma 259
Author Biography
Abstract The increasing rate of cyberattacks based on the DDoS principle has
created various new areas of concern for information security. It has also raised a
pertinent question—Are we protected against such attacks? With significant rise in
the number of attacks and resulting reports of high vulnerability to ICMP flood
attacks, perhaps we need to reconsider and revisit the pros and cons of the ICMP
protocol. In this paper, we mainly focus on giving readers a brief outline of DDoS
attacks and its constituents, primarily the ICMP protocol. We also present a survey
and the research findings that show the rising vulnerability to ICMP and subse-
quently DDoS.
1 Introduction
V. Chauhan (&)
Knowledge Graph Department, Binary Semantics Pvt. Ltd., Gurgaon, India
e-mail: [email protected]
P. Saini
Department of Information Technology, Bharati Vidyapeeth’s College of Engineering,
GGSIPU, New Delhi, India
e-mail: [email protected]
Fig. 1 A geography of who’s hitting whom in denial-of-service cyberattacks. The large unknown
reflects the difficulty of measuring the threat [7]
ICMP Flood Attacks: A Vulnerability Analysis 263
2 Related Work
The Internet Control Message Protocol (ICMP) is the means to give feedback about
network troubles that is thwarting delivery of packets throughout the network.
Upper protocols, like TCP, are able to understand that packets are not getting
delivered, but ICMP provides a method for discerning more catastrophic problems,
such as “TTL exceeded” and “need more fragments.”
The ICMP protocol is used for sending various messages to convey network
conditions. The majority of ICMP message types are necessary for proper operation
of TCP, IP, and other such protocols. ICMP is not evil and should not be blocked.
An ICMP flood is said to have happened when an attacker makes use of a botnet to
send large amounts of ICMP packets to the target server in an attempt to exhaust
any available bandwidth and prevent access to the legitimate users. This attack is
considered “successful” when a huge number of sources are able to send sufficient
ICMP traffic so as to consume and exhaust all available bandwidth of the victim’s
network.
One instance of this attack is the “ping” command. The “ping” command is
primarily used to test network connectivity by checking whether your device is able
to send and receive data to/from other device in the network, i.e., between two
264 V. Chauhan and P. Saini
points in a network. However, this command can be given with different variables
to make the ping larger in size and occur more often. Efficient application of such
parameters and with adequate source systems initiating traffic will finally lead to the
utilization of available system bandwidth (Fig. 2).
Everybody likes ping. It is simple. It is useful. And it also does precisely what the
sonar-inspired name indicates.
Ping tells you if a remote computer is responding to network requests.
The ping utility was written by Mike Muuss, a senior scientist at the U.S. Army
Research Laboratory [5]. It makes use of IP/ICMP ECHO_REQUEST and
ECHO_REPLY timed packets in order to probe the “distance” to the target machine
(Fig. 3).
ICMP Flood Attacks: A Vulnerability Analysis 265
4 Vulnerability Analysis
A survey was conducted wherein data was collected about various websites
vis-a-vis their vulnerability to ICMP.
Major categories of websites targeted:
• Social Networking Websites
• E-Commerce Websites
• Email Service Providers
• Search Engines
• Government Websites
• Private Enterprises
All the website and portals in these categories are privy to sensitive information
and act as prime targets for attackers and hackers alike.
Step 1. A list of 100 top-most websites from each category was prepared and a
database was maintained using MS-Excel.
Step 2. A range of parameters was decided, based on which results would be made
and conclusions drawn. The parameters were as follows:
• Destination IP Address
• TTL, i.e., Time To Live
266 V. Chauhan and P. Saini
The results of the survey and the conclusions drawn are given below (by category):
1. Social Networking Websites: Approximately, 20% of websites have blocked
ICMP requests and are therefore not vulnerable to such attacks. The remaining
80% of these though have allowed ICMP and are therefore vulnerable to such
attacks. Some of these are namely, Facebook, Twitter, and LinkedIn which are
privy to a huge amount of sensitive information.
2. E-Commerce Websites: Around 30% of such websites rejected ICMP echo and
reply requests and are not vulnerable to such attacks. The remaining 70%
though is highly vulnerable and can be brought down by such attacks in the
future. These include sites like Flipkart, Snapdeal, and Jabong which are among
the top e-commerce sites in India.
3. Email Service Providers: Here, we have a 50:50 ratio. 50% of such email
providers are vulnerable to ICMP-based DDoS attacks. Hackers can gain sen-
sitive knowledge that can lead to expensive losses. These include Gmail,
Outlook, Hotmail, etc.
4. Search Engines: Only 30% of search engines have blocked ICMP requests,
while the remaining 70% is highly vulnerable and sensitive to such attacks.
These search engines can be manipulated to divert traffic which happened in the
case of GitHub attack.
5. Government Websites: More than 90% of such websites that includes Indian
and US Government websites are aware of such possibilities and have subse-
quently blocked an ICMP request on their servers and sites. The rest though are
still highly vulnerable. These include the US Air Force and NASA.
6. Private Enterprises: 40% of private enterprises have blocked ICMP requests
and hence are protected against this type of attack. The other 60% have allowed
such requests and hence are vulnerable. These include Apple Inc., General
Electric, Exxon Mobil, etc.
From the above results, we can conclude that around 47% of all website are
vulnerable to such attacks. Therefore, these types of attacks need to be mitigated.
5 Conclusions
Through this paper, the most crucial terms with regard to DDoS and ICMP and the
association between these terms have been explained. It has been shown—in par-
ticular regarding cybersecurity—that ICMP is a great hacking tool. Engineers,
administrators, security officers, etc. need to be aware of the dangers. The data
present within these messages can be misused by attackers to exploit known vul-
nerabilities in the system. We have seen all through this paper that ICMP can and
has been used in many stages of an attacker’s progress in a system compromise. We
have also seen that ICMP is not just being used in the reconnaissance and probing
phase which is most understood but it has also been used for exploiting systems as
well as in certain occurrences as a covert passage for attacker’s communication.
Further, research should seek to test this vulnerability which has not been shown by
this study. Practical implementation and testing can be done to further narrow down
268 V. Chauhan and P. Saini
References
1. Website DDoS protection. Stop DDoS attacks against your website! https://sucuri.net/website-
firewall/ddos-protection
2. Douligeris C, Mitrokotsa A (2003) DDoS attacks and defense mechanisms: classification and
state-of-the-art, Greece
3. Eden L The truth about ICMP, global information assurance certification paper, SANS Institute
4. Myers L (2014) Guide to DDoS attacks, integrated intelligence center technical white paper,
Center for Internet Security
5. Atwood J (2007) The story about PING. http://blog.codinghorror.com/the-story-about-ping/
6. ICMP attacks illustrated, SANS Institute InfoSec Reading Room
7. “The Internet’s Aswarm in Denial of Service Attacks and It’s Getting Worse.” (2014) June
2014. http://www.forbes.com/sites/bruceupbin/2014/06/18/were-aswarmin-denial-of-service-
attacks-and-its-getting-worse/
Statistical Approach Using Meta
Features for Android Malware
Detection System
1 Introduction
Smartphones installed with Android has gained its popularity over its counterparts
like iOS, Blackberry, Symbian, and Windows. Due to increased popularity of
Android devices, malware writers and hackers have found interest in identifying
vulnerabilities and compromising these devices. This has given birth to malicious
apps in Google Play Store and in many third-party app stores. Traditional
signature-based systems [1] have found to be incapable in the identification of
zero-day malware. Exponential increase in growth of malware being repackaged
with legitimate apps [2] has raised serious security concern. Many prior works in
the domain of detection of malware are considered using single feature such as
permissions, API, opcodes, etc. for identifying suspicious file. As such these
methods did not scale in identifying unseen sample, there was a demand felt in
the detection of files by composing different types of features in other words
composing diverse optimal feature vectors. The main contribution of work
includes the following: (a) Implementation of a malware scanner by incorporating
static analysis using machine learning approach; (b) Classification of Android
malware against benign with less classification overhead in minimum time;
(c) High performance is achieved due to the application of GSS (Galavotti–
Sebastiani–Simi) and mutual information feature selection methods; and
(d) Composite (hereafter meta referred to as composite) feature space model
results in high F-measure of 0.996 in 1.01 s. The remaining section is organized
as follows: Sect. 2 introduces related work. Proposed framework is discussed in
Sect. 3. Experimental results are covered in Sects. 4, and 5 concludes the paper
with scope for future enhancement.
2 Related Work
In [3], feature sets from manifest file and dex code of different apps on Drebin
dataset were analyzed. Authors in [4] rank permissions and determine a subset of
critical permissions using sequential forward selection and PCA. In [5], proba-
bilistic generative models ranging from Naïve Bayes to advanced hierarchical
mixture models were utilized for scoring permissions. Proposed approach catego-
rizes an app based on the usage of permissions, the function category of an app, and
permissions requested by other apps in [6] to risky report if an app is malicious.
A static analysis method is proposed in [7] for Android malware detection by
extracting creator information from every app. MAST (Mobile Application Security
Triage) architecture was proposed by authors in [8] to exhibit malicious behavior.
In [9], Andromaly was proposed that monitors smartphone features and events
contributing to 88 features to detect malware, whereas in [10] to detect mali-
ciousness, Kirin looks for app permission and Stowaway in [11] detect overprivi-
leged apps by analyzing API calls. Droidmat in [12] extracts API calls, intents, and
permission to detect malware. API calls are extracted from decompiled source code
for identifying malicious app in [13].
Statistical Approach Using Meta Features for Android Malware … 271
3 Methodology
Malware samples are collected from Drebin dataset (5560 samples in 6 folders) in
which three folders Drebin-0, Drebin-1, and Drebin-2 with a total of 3000 malware
samples and 1631 benign samples downloaded from Google Play Store are used for
experimentation. All input.apk files are disassembled using Androguard tool [15] to
convert Android manifest file into human readable.xml format later used for feature
extraction. The malware and benign samples are randomly divided into train set and
test set in the ratio 60:40 resulting in three datasets for training and three for testing.
The static features extracted from samples are depicted in Table 1.
Table 1 Static features extracted from android manifest file of various applications
Static features Descriptions Examples
Hardware Contains requested hardware components CAMERA.FLASH, GPS,
components WI-FI
Requested Granted by user at the time of installation SEND_SMS,READ_LOGS,
permissions WRITE_GMAIL
App Four components: activities, services, PACKAGE_INSTALL,
components providers, and receivers DOWNLOAD_HIDE
Filtered Contains action and category components BOOTSERVICE,
intents USER_PRESENT
272 M. M. John and P. Vinod
Initially, from each feature set, rare attributes that are not relevant to be used for
developing learned models are removed using Naïve Bayes approach. A unique list
with feature, document frequency, and term frequency is used to eliminate irrele-
vant attributes using Eq. (1).
where Ci denotes class [Benign (B) or Malware (M)] and X denotes a feature. Naïve
Bayes score is calculated for each feature and is sorted in descending order. This
would preserve the attribute that identifies a target class. Then, features with Naïve
Bayes score greater than zero are extracted resulting in two feature lists: pruned
malware and benign list of features. This initial pruning is repeated for all categories
of feature from.apk samples and the pruned set is further given to feature selection
techniques to obtain relevant features with minimum redundancy that is used for
constructing learning models.
where tj represents feature and ck represents class (B, M). P tj ; ck and P tj ; ck
represent joint probability of the presence
and absence of a feature in a particular
class [say ck ¼ M (Malware)]. P tj ; ck and P tj ; ck represent joint probability of
the presence and absence of a feature in alternate class [say ck ¼ B (Benign)]. The
features are sorted in descending order of GSS score calculated using (2), and these
significant attributes are involved in the preparation of malware and benign models.
Statistical Approach Using Meta Features for Android Malware … 273
1 Pðf ; MÞ
MIðf ; MÞ ¼ log ð3Þ
PðMÞ Pð f Þ PðMÞ
1 Pðf ; BÞ
MIðf ; BÞ ¼ log ð4Þ
PðBÞ Pð f Þ PðBÞ
Larger MI
Fmax ¼ ð5Þ
Smaller MI
Fmax is calculated using (5) and the result is sorted in descending order and
subsequently used for preparing training models.
Independent training models are constructed for features like action, activity, cat-
egory, hardware components, requested permissions, content providers, broadcast
receivers, and services. Each instance is represented in the form of a Boolean
vector, where 1 denotes the presence and 0 denotes the absence of feature in the
sample. The models are generated using three classification algorithms imple-
mented in WEKA [18]. They are SVM [18] with linear, polynomial, radial, and
sigmoid kernels; Random forest [18]; and rotation forest [18]. The accuracy, false
positive rate (FPR), true positive rate (TPR), F-measure, and area under ROC
(Receiver Operating Characteristic) curve (AUC) are measured at variable feature
length during training phase. The model at optimal feature length is used to test the
unknown samples. Subsequently, different category features of each optimal model
are aggregated to form composite feature space, and this learning model is used to
identify new samples that are not used in modeling.
Malware specimens are collected from Drebin dataset and benign samples from
Google Play Store. Experiments were conducted on Ubuntu 14.04 platform with
8 GB RAM. The performance of the proposed framework is estimated using
Eqs. 6–10.
274 M. M. John and P. Vinod
2ðPrecision RecallÞ
F-MeasureðF Þ ¼ ð6Þ
ðPrecision þ RecallÞ
TP
PrecisionðPÞ ¼ ð7Þ
TP þ FP
TP
RecallðRÞ ¼ ð8Þ
TP þ FN
TP þ TN
Accuracyð AÞ ¼ ð9Þ
TP þ FN + TN þ FP
4.1 Results
The effect of F-measure on independent models with linear SVM (L), random forest
(Rdm), and rotation forest (Rtn) classifiers with GSS are shown in Figs. 2, 3, 4, 5, 6
and with mutual information in Figs. 7, 8, 9, 10, 11. D-0, D-1, and D-2 represent
dataset-0, dataset-1, and dataset-2, respectively.
In Figs. 2. 3, 4, 5, 6, 7, 8, 9, 10, and 11, five independent training models out of
eight models discussed in model generation and prediction using GSS and MI
shows good results and is plotted. The results of linear SVM and random forest
Fig. 4 Performance
considering permission using
GSS
classifiers are better than that of rotation forest. Malware model of GSS plotted
above outperformed benign GSS model. In Fig. 2, GSS learning malware model of
action using 236 features pruned out of 2727 features (236/2727), activity (1350/
13,169) with linear SVM in Fig. 3 and permission (130/562) trained using random
forest (refer Fig. 4) shows high F-measure of 0.793, 0.977, and 0.979. Receiver
(390/2063) and service (340/1861) learning model depicted in Figs. 5 and 6
276 M. M. John and P. Vinod
Fig. 6 Performance
evaluated with service
attributes using GSS
Fig. 10 Evaluate
performance with receiver
attributes using MI
Table 2 Prediction of new samples for independent features in GSS using AUC
Linear SVM Random forest Rotation forest
Features D0 D1 D2 D0 D1 D2 D0 D1 D2
Action 0.797 0.801 0.794 0.924 0.921 0.921 0.835 0.839 0.840
Activity 0.973 0.979 0.974 0.984 0.995 0.988 0.790 0.816 0.776
Permission 0.910 0.925 0.924 0.994 0.996 0.996 0.984 0.986 0.980
Receiver 0.875 0.875 0.872 0.891 0.890 0.888 0.729 0.739 0.735
Service 0.900 0.891 0.876 0.904 0.894 0.880 0.731 0.720 0.745
5 Conclusion
References
1. Grace MC, Zhou Y, Zhang Q, Zou S, Jiang X (2012) Riskranker: scalable and accurate
zero-day android malware detection. In: MobiSys, pp 281–294. ACM
2. Zhou W, Zhou Y, Jiang X, Ning P (2012) Detecting repackaged smartphone applications in
third-party android marketplaces. In: Proceedings of second ACM conference on data and
application security and privacy, pp 317–326. ACM
3. Arp D, Spreitzenbarth M, Hubner M, Gascon H, Rieck K (2014) Drebin: effective and
explainable detection of android malware in your pocket. In: NDSS. The Internet Society
4. Wang W, Wang X, Feng D, Liu J, Han Z, Zhang X (2014) Exploring permission-induced risk
in android applications for malicious application detection. IEEE Trans Inf Forensics Secur 9
(11):1869–1882
5. Peng H, Gates CS, Sarma BP, Li N, Qi Y, Potharaju R, NitaRotaru C, Molloy I (2012) Using
probabilistic generative models for ranking risks of android apps. In: ACM conference on
computer and communications security, pp 241–252. ACM
6. Sarma BP, Li N, Gates CS, Potharaju R, Nita-Rotaru C, Molloy I (2012) Android
permissions: a perspective combining risks and benefits. In: SACMAT, pp 13–22. ACM
7. Kang H, Jang J, Mohaisen A, Kim HK (2015) Detecting and classifying android malware
using static analysis along with creator information. Int J Distrib Sens Netw, 479174:9
8. Chakradeo S, Reaves B, Traynor P, Enck W (2013) MAST: triage for market-scale mobile
malware analysis. In: Proceedings of security and privacy in wireless and mobile networks,
ACM
9. Shabtai A, Kanonov U, Elovici Y, Glezer C, Weiss Y (2012) Andromaly: a behavioral
malware detection framework for android devices. J Intell Inf Syst 38(1):161–190
10. Enck W, Ongtang M, McDaniel PD (2009) On lightweight mobile phone application
certification. In: ACM conference on computer and communications security, pp 235–245.
ACM
11. Felt AP, Chin E, Hanna S, Song D, Wagner D (2011) Android permissions demystified. In:
ACM conference on computer and communications security, pp 627–638. ACM
12. Wu D, Mao C, Wei T, Lee H, Wu K (2012) Droidmat: android malware detection through
manifest and API calls tracing. In: AsiaJCIS, pp 62–69. IEEE
13. Cen L, Gates C, Si L, Li N (2013) A probabilistic discriminant model for android malware
detection with decompiled code. In: Dependable and secure computing, IEEE
14. Drebin Dataset. http://user.cs.uni-goettingen.de/*darp/drebin/
15. Androguard. http://code.google.com/p/androguard/
16. Largeron C, Moulin C, Gry M (2011) Entropy based feature selection for text categorization.
In: SAC, pp 924–928. ACM
17. Sebastiani F (2002) Machine learning in automated text categorization. ACM Comput Surv
34:147
18. Weka. http://www.cs.waikato.ac.nz/ml/weka
Composite Email Features for Spam
Identification
1 Introduction
Email is one of the most popular means of communication in the era of Internet.
Spam, referred to as unsolicited commercial email (UCE) or unsolicited bulk email
(UBE) [1], consumes most of the bandwidth. Moreover, it can also quickly con-
sume server storage space. It is observed that the nature and the characteristics of
spams change over time [2] and this demanded efficient approach for filtering
unwanted emails. There are many techniques designed and developed to categorize
emails. All these methods look for some known patterns or features (words) alone
that usually appear in spam or ham messages, to classify the emails. These methods
do not consider the syntactic and the semantic peculiarities of the messages. This
was the primary motivating factor to discover varied features existing in the spam
emails. Also, the studies done on author gender identification [3] by applying NLP
[4, 5] became another inspiration behind this proposed work to investigate the
contribution of tag-based features and other linguistic attributes in developing an
email spam classification model with minimum FPR. Contributions of our approach
are (a) prepared an efficient model for classification of spam and ham mails,
(b) higher performance is obtained by applying feature selection method, (c) eval-
uated efficiency of each category of attribute set in spam categorization, and (d) an
overall performance, i.e., F-measure of 0.9935 with a smaller FPR of 0.0004 jus-
tifying the applicability of our proposed approach in real-time spam filtering
system.
The reminder of the paper is organized as follows. In Sect. 2, a review of the
related works is done. The proposed mechanism is described in Sect. 3. Section 4
discusses details of experiments and the results of the study. Finally, inferences are
included in Sects. 5 and 6 presents the concluding remarks of the study.
2 Related Works
In [3], an author gender identification technique was proposed and it could achieve
accuracy of 85.1%. A new one-class ensemble scheme is put forward, which uses
meta-learning to combine one-class classifiers in [6]. Blanzieri and Bryl [7] have
discussed various machine learning applications for email spam filtering. Menahem
et al. [8] implemented a new sender reputation mechanism based on an aggregated
historical dataset. In [9], the authors designed a fusion algorithm based on online
learners and experimented on TREC (Text REtrieval Conference) and other data-
sets. Comprehensive review on machine learning approaches to spam filtering is
discussed in [10]. Drucker et al. [11] investigated the applicability of Support
Vector Machines (SVMs) in classifying email as spam or legitimate mail.
Three-layer Backpropagation Neural Network (BPNN) technique is implemented
on datasets PU1 and Ling, resulting in 97 and 99% of classification accuracy with
less execution time [12]. A three-way decision approach (accept or reject or further
exam) is discussed and experiments on SpamBase dataset resulted in reduced
misclassification rate in [13]. Wu [14] utilized spamming behaviors with a back-
propagation neural network, employed on datasets from Hopkins, Reeber, etc. to
achieve improved performance (FPR = 0.0063).
Composite Email Features for Spam Identification 283
3 Proposed Methodology
Email spam detection process is carried out through different steps (refer Fig. 1) and
evaluated over Enron-spam dataset [15, 16]. The following subsections introduce
the proposed approach.
Email body is extracted from each email in the dataset. The resulted collection of
extracted email body is partitioned into train and test (60:40 ratio). Style markers
are treated as features in our approach. There are 31 characters, 38 words, 35 tags, 3
structural features, and 10,280 Bag-of-words extracted from mail body.
• Character-based features [3] include total number of characters (C), ratio of
total number of lower case letters (a–z) and C, ratio of total number of uppercase
characters and C, fraction of total number of digital characters and C, fraction of
total number of white-space characters and C, ratio of total number of tab space
characters and C, and fraction of number of special characters and C (25 special
symbol features).
• Word-based features [3] consist of total number of words (N), average length
per word, ratio of total different words and N, fraction of words longer than 6
characters and N, ratio of total number of short words (1–3 characters) and N,
Guirad’s R, Herdan’s C, Rubet’s K, Maa’s A, Dugasts U, L. Janenkov and
Neistoj Measure, Sichel’s S, Yule’s K measure, Simpson’s D measure, Hapax
Dislegomena, Hapax legomena, Honore’s R measure, Entropy, and ratio of
word length frequency distribution and N.
• Function words [3] (or grammatical words or tag-based features) are words that
express grammatical relationships with other words within a sentence. Tags are
extracted from email text using NLTK (Natural Language Tool Kit) [17] in
python, and Part-of-Speech (POS) [18] tagging is done using Penn Treebank [5]
tag set.
• Structural features [3] represent the way an author organizes the layout of a
message. The main features are total number of lines, total number of sentences
(S), and average number of words per sentence.
• In Bag-of-Words, all sentences in each email body are tokenized into a set of
words and frequency of every term is counted within each file (called as term
frequency).
Feature selection determines optimal attributes from a huge attribute space without
changing physical meaning of the attribute. The main benefits of dimensionality
reduction (or feature selection) are (a) elimination of redundant features, (b) re-
duction in noise thereby increases accuracy of classifiers, (c) reduction of time
complexity of classification, and (d) minimization of over-fitting of the training
data.
A weighting method called TF–IDF–CF [19] is applied in our proposed
approach. This method is developed based on TF–IDF (Term Frequency–Inverse
Document Frequency). It says that if a term appears in more documents, then it
becomes less important, and the weighting will also be less. A new attribute, called
as class frequency, is introduced to assess the frequency of each term in every
document within a specific class. A general form of TF–IDF–CF is shown in
Eq. (1).
In Eq. (1), tfij indicates the term frequency of term j in document i, N is the total
number of instances in the dataset, and nj indicates the number of documents that
term j occurs. The term ncij represents the number of files within the same class
c where document i belongs to and term j appears, Nci gives the total count of
documents within the same class c where document i belongs to. The algorithm for
extracting significant words is given below.
Composite Email Features for Spam Identification 285
Feature selection produces a reduced feature vector table (FVT) which is taken as
the input for training the classifiers. Multinomial Naïve Bayes (MNB) and support
vector machine are used as classifiers in this investigation. Individual training
models are created for each category of feature during training phase. The model
with highest F-measure is chosen for prediction. Finally, the optimal models
obtained from each category of features are aggregated to develop a composite
feature space used for building spam and ham model, subsequently used for
prediction.
The experiment was performed on Ubuntu 14.04 platform with the support of Intel
core 7 and 8 GB RAM. In this work, 12,045 ham and 4496 spam emails have been
chosen from Enron-spam dataset. The classification models are generated by
LibSVM (kernels k0 (Linear), k1 (Polynomial), k2 (Radial), and k3 (Sigmoid)), and
Multinomial Naïve Bayes (MNB) in WEKA [20]. When a ham is misclassified as
spam, a false positive (FP) occurs. If ham data is predicted as ham then it is known
as true negative (TN), whereas if spam is correctly classified as spam data then it is
true positive (TP). When a spam is wrongly taken as ham, it is considered as false
negative (FN) [21, 22]. In this analysis, F-measure (also called as F1-score) and
286 P. George and P. Vinod
FPR are used as the significant evaluation parameters. The F1-score can be inter-
preted as a weighted average of the precision and recall, and it ranges from 0 to 1.
Precision (P) is a measure of the accuracy provided that a specific class has been
predicted. Recall (R) measures the proportion of actual positives which are correctly
identified as such.
5 Inferences
It has been analyzed that SVM classifier performs well with large number of
features, but it is computationally expensive. Performance is observed higher when
stop words are removed from the text before model construction. Independent style
markers with small number of features produced insignificant results in terms of
F-measure, which is clearly visible for character-based, word-based, and structural
features. Hence, these attributes are not sufficient enough to prepare an efficient
spam filtering model independently. This is due to the absence of attributes having
high correlation with target class. As the features in the feature space increase, the
performance also improves, since relevant attributes contributing to the effective
classification appear as a candidate in the optimal feature space. This is why
bag-of-words produced a highest F-measure with larger feature space of size
10,153. Tag-based attributes and bag-of-words played an important role in the
generation of composite model as they could produce lower FPR value and an
288 P. George and P. Vinod
6 Conclusion
References
Abstract This paper debates about the role of multiple encryptions in biometric
devices with a particular focus on the privacy and security benefits of biometric
devices. This research paper is proposed to engage a larger number of data users to
consider the beneficial role of multiphase encryption in biometric devices with
enhanced security solutions. This paper discusses about how multiphase encryption
can be promoted with biometric devices in order to overcome the present loopholes
of the security devices. In current scenario, wireless security demands to provide an
approach for securely verifying the user’s identity, authenticating the data access
and certifying the security applications. The security of data has nowadays become
challenging issues that comprises areas like data encryption, protected communi-
cation channel, and reliable third party to preserve the databases. The immediate
growth in the area of information technology, the super-secure communication of
confidential data is highly required. Biometrics technology has been extensively
used in user’s verification and identification, but there are several security issues to
provide adequate security. Therefore, using multiple encryption techniques in
biometric devices, we can enhance the data security enormously.
1 Introduction
1.1 Biometrics
traits for verification purposes. As the matter of fact, everyone in the world is
unique and hence this uniqueness can be used for identity verification. In simple
language, it can be stated that biometric technology is typically considered to
examine human characteristics for identification and security purposes. The most
common parameters which are measured under this technology are fingerprints,
hand, eye, face, and voice. The fingerprints’ identification systems have been
installed as access control systems since 1960s. The biometric product based on the
geometry of the hand was introduced during the 1970s in many access control
applications. Eventually, the systems using biometric technology were enhanced
and moved from the geometry of the hand to the characteristics of the eye. In the
middle of 1980s, the system was executed to analyze the unique patterns of retina
while parallel improvements were on to recognize the iris patterns.
that is, his eye, voice, fingerprint, or any biometric used for the security purpose.
Then, an image of this metric is generated and a key as well. Both the key and the
image are bind together and stored in the database. After this template has been
saved, the key and the image are just destroyed. If the user wants to access its data
from the database, again the biometric input will be checked with saved template. If
the image matches, then the key is retrieved and the image is discarded. Now, the
user can access the particular application as per his requirement. If the image does
not match, the request is rejected [4, 5].
294 H. Gupta and C. A. A. Anne-Marie
2 Proposed Model
In the implementation of the enrolment phase, two different ciphertexts are used.
For producing image by the sensor at the time of the enrolment phase, two different
ciphers are used. The image produced at the enrollment time is duplicated. A copy
is binded with a cryptographic key and a first biometric encrypted template is
produced. A pseudo-random number (in bits format) is now generated and this
activates the production of a second independent key. The second key is used to
bind the second copy of the image and the first biometric encrypted template
together. And the final encrypted template is produced which is finally stored as
multiple encrypted data. The keys and the images were generated initially and
discarded once the final image is developed and stored.
When the user wants to access a desired application, two images are taken from the
sensor device. The first image used to decrypt the first biometric encrypted image,
and the second copy of the biometric image is used for verification purpose. The
subsequent keys are retrieved and the user can access the application. If one phase
fails, the process is rejected and access is denied (Fig. 2).
3 Conceptual Framework
In the enrollment phase, we divide the inner work of proposed multiple encryptions
into three stages:
In the first stage: P1 (a)
When the user wants to log in the first time to a particular application, he just
registers with his biometric input with the help of a sensor. Suppose the biometric
input is the fingerprint. The image of the fingerprint is taken from the sensor, and
Role of Multiple Encryptions in Biometric Devices 295
then duplicate copy is generated and saved before being converted into binary
format. Each sample of fingerprint image is called, respectively, as F1(n) and
F2(n) where n is the metric. Then, a pseudo-random number Key(A) is generated
and taken as a cryptographic key. This cryptographic key is a single digit between
0 and 9 that is converted into binary number and then in octal format.
The first function F1(n) is matricized into matrix of (m m) size where maxi-
mum value of m = 2. Those matrices will be generated sequentially according to
number of binary digits in a line and a column. If there is no enough digit for
making a matrix, then the size of the matrix will be padded with NULL value (0),
which should create a very less distortion (almost null according to the correlation
algorithm).
The next step is doing a matrix multiplication between each and every matrices
and the same Key(A). Results will be written sequentially according to initial
position of matrix. The output from phase 1 is called as biometric encrypted
template as BE(1) which is the first template and is binded with Key(A). One copy
is stored in database for retrieval purpose. Now, the Key(A) input is discarded.
The output of the first part of stage P(1) as BE(1) will be used as the input of the
stage P1(b). The second function that is nothing but the digitized and binary version
of the input image F2(n) is matricized into matrix of (m m) size where maximum
value of m = 2. The same process as above is followed. Then, BE(1) is multiplied
296 H. Gupta and C. A. A. Anne-Marie
with each and every matrices produced with F2(n) as F2(n): (m m) * BE(1) =
BE(2). The output is BE(2). The BE(1) input along with the function F2(n) is
discarded. One copy of BE(2) is stored in database.
In the second phase: P2
A pseudo-random number Key(B) is generated and taken as a cryptographic key.
This cryptographic key is a single digit between 0 and 9 that is converted into
binary number of 1 octet.
In this phase, output BE(2) is EXORed with the cryptographic key Key(B)
(Fig. 3).
The output is a biometric encrypted template BE(3) which is bind with the
cryptographic key, Key(B). Then, BE(3) is stored in database and key(B) is
discarded.
In the last phase: P3
The third pseudo-random key, Key(C), is generated. In this encryption phase, an
idea from Ceaser cipher technique of encryption will be followed.
The cryptographic key is taken as a three-digit octal number, and its value will
predict the position to which it should be added. Suppose the PRNG number is 3
and all numbers will be grouped into three digits. And 3 in octal system is 3 = 011.
It means, 011 will be added to each 3th forward position in the list of numbers in
BE(3). If in the first row, we have
BEð3Þ : 1 0 1 1 0 1 0 1 1 1 1 0 1 0 1 0 1 1 0 0 0 1 1 1 01 0 1 1 0 . . .
1 0 1 1 0 1 0 1 1 1 1 0 1 0 1 0 1 1 0 0 0 1 1 1 0 10 1 1 0
To the 3th binary octet, then the 6th, then the 9th in a raw, and so on, we will add
011.
So BE(3) + key(C) = 1 0 1 1 0 1 (0 1 1 + 0 1 1) 1 1 0 1 0 1 (0 1 0 + 0 1 1) 0 0 0
1 1 1 (0 1 0 + 0 1 1) 1 1 0… = 1 0 1 1 0 1 1 1 0 1 1 0 1 0 1 1 0 1 0 0 0 1 1 1 1 0 1 1
1 0… = BE(4)
The result is biometric encrypted template BE(4) which stored in database, bind
with cryptographic key Key(C) and linked to the intended application. Key(C) is
then discarded.
The retrieval phase is a combination of correlation algorithm, that is, the basic
algorithm in biometric encryption and also, some mathematical operations taken in
a block as multiple encryption different levels. The objective of the retrieval phase
is to allow the user to access a particular application after authentication and ver-
ification (Fig. 4).
At the retrieval phase, the biometric is taken as an input from the sensor. And
this biometric image is digitized, binarized, and then every single bit is compared
with F1(n) copy that was stored in the database. Approximately, 99% match can be
tolerated. If there is a match then the process is pursued, if not, the access is denied.
Then, the Key(A) is released and input image A(n) is matricized as F1(n) and
multiplied with the cryptographic key Key(A). The result is called as BEr(1) and it
is compared with BE(1); if there is a 99% match, then A(n) is discarded and go to
the next phase. If not, process is rejected.
F1(n) is now multiplied with BEr(1) and result is BEr(2). Then, BEr(2) is
compared with BE(2). If there is a 99% match, then Key(B) is released, and BEr(1)
is discarded. If not, the process is rejected. BEr(2) is EXORed with the key,
Key(B), and produces a biometric encrypted BEr(3) sample. Then, BEr(3) is
compared with BE(3).
If there is a 99% match, then Key(C) is released, and BEr(2) is discarded. If not,
the process is rejected. BEr(3) is now Ceaser Ciphered with the cryptographic key
Key(C) like in the enrollment phase. The result is called as BEr(4) and it is
298 H. Gupta and C. A. A. Anne-Marie
compared with BE(4); if there is a 99% match, then the link to the application is
activated and the user gets access to that application. After that, BEr(4) is discarded.
If there is no match, the process is interrupted.
Role of Multiple Encryptions in Biometric Devices 299
4 Future Work
After completion of this research paper, we came across some limitations such as
• Level of security implemented,
• Ease of application access,
• Availability of resources to implement the technology on a common platform,
and
• Time management.
We are planing to work more effectively on the prior aspects of the biometric
technology like enhancing more security. This research paper is only a theoretical
one and we are not 100% sure that everything that we have proposed will effec-
tively work in biometric devices. Everything proposed here will be implemented at
software level. So the next step will be to create an application based on all
above-given mathematical calculations.
5 Conclusion
References
Author Biographies
Dr. Himanshu Gupta is working as a Senior Faculty Member in the well reputed Indian
university Amity University Uttar Pradesh, Noida. He completed all his academic as well as
professional education from reputed central university Aligarh Muslim University, Aligarh (Uttar
Pradesh) India. He has visited Malaysia, Singapore, Thailand, Cambodia, Vietnam, Indonesia,
Hong Kong, Macau and China for his academic and research work. He has delivered many
Technical Sessions on “Network Security & Cryptography” in the field of Information Technology
in various reputed International Conferences, World Summit and other foreign universities as an
Invited Speaker. He has more than 60 Research Papers and Articles in the field of Information
Technology, which have been published in various reputed Conference Proceedings and Journals.
C. Aka Assoua Anne-Marie is associated with Amity University, Noida as a research student and
having her expertise in Network Technology & Management. She has been associated in many
academic and research activities in the area of Network Technologies. She earned the CCNA
Certification during her stay as a research student in the Amity University, Noida.
Buffer Overflow and SQL Injection:
To Remotely Attack and Access
Information
1 Introduction
In the electronic world, security is the major issue on the internet, intranet, and
extranet. Ethical hacking is a term which is used to increase security by identifying
and overcoming those vulnerabilities on the systems owned by third party. Attacker
uses vulnerabilities as an opportunity to attack software and web application. Thus,
system needs to be protected from attacker so that attacker cannot hack information
and make it misbehave according to him/her. So, ethical hacking is a way to test
and to identify an information technology environment for present vulnerabilities.
Software is used everywhere in the digital world. But due to the flaws in soft-
ware, software fails and attacker takes this as an advantage and uses this oppor-
tunity to make software misbehave and use according to them. Flaws increase the
risk to security. Some of the software developer manages the software risk by
increasing the complexity of the code, but absolute security cannot be achieved.
According to the literature survey, some of the software flaws which lead to security
vulnerabilities are Buffer Overflow (BO), Incomplete Mediation (IM), and Race
Condition (RC) [1].
The other class of vulnerabilities that exist in the web application can be
exploited through SQL injection. Attacker takes advantage of the unintentional
flaws in the input validation logic of Web components. SQL injection attack leads
to high security risk to the web applications which allow attackers to access
databases completely. These databases contain user information and if this infor-
mation is accessed by the attacker, the confidentiality of the user will be leaked and
thefts and frauds can take place. In many cases, attackers use an SQL injection
vulnerability to take full control of web application and corrupt the system that
hosts the Web application.
Buffer overflow and SQL injections are some of the vulnerabilities in the soft-
ware and the web application, respectively, which are discussed in detail in this
paper.
This paper gives an overview of one of the flaws that exist in the software, i.e.,
buffer overflow and how this flaw leads to the security vulnerabilities that can be
exploited and what are the preventives measures that can be taken to protect it from
the attackers. This paper also explains one of the famous attacker’s techniques to
access information from the database of the web application using Kali Linux.
The SQL queries in Kali Linux that are used to retrieve information from the
database of a web application are shown for a particular website. This paper pro-
vides description and example with screenshot of how these attacks can be per-
formed and what will be the outcome.
2 Buffer Overflow
(a) Denial-of-Service Attack—Buffer overflow flaw may likely cause system crash,
so attacker exploits this vulnerability to launch denial-of-service attack.
(b) Inject Attack Code—Attacker can manipulate the code to
(i) Overwrite the system data.
(ii) Overwrite the data in the memory in such a way that it transfers the code
to malicious code, i.e., pointer points to injected malicious code. Buffer
overflow vulnerabilities mostly dominate in the class of remote penetra-
tion attack.
Figure 1 shows the structure of memory organization of CPU. Here, text stores
the code of the program, the data section consists of text and static variables, heap
stores dynamic data, and stack section (shown in Fig. 2) stores the dynamic local
variables, parameters of the functions, return address of the function call (where the
control will be transferred after the function executes), stack pointer points to the
top of the stack. Stack grows from high address to low address (while buffer grows
from low address to high address).
The main goal of the attacker is to take buffer overflow as an advantage and to fetch
the control of the privilege program by subverting the function of that program.
Attacker tries to attack the root program and execute code similar to shellcode. To
achieve this goal, two sub-goals need to be achieved [3].
(a) To alter the victim’s program by making it to jump to random memory location,
with suitable parameter loaded into register and memory.
(b) To alter victim’s program by adding malicious code to victim’s program
address space to jump to address where malicious code is injected.
Attacker overwrites program with arbitrary sequence of byte with goal of corrupting
the victim program. It is done by making the victim’s pointer to point to random
address.
According to Table 2, if attacker tries to use more memory (>10), buffer over-
flow will overflow into the space where the return address is located. Attacker can
overwrite this return address with the random bits or random address; by this, the
program will jump to random memory location after function execution [4] and
may lead to program crash as shown in Fig. 3b.
Buffer Overflow and SQL Injection … 305
Stack and heap are two areas of memory that a program used for reading and
writing, i.e., buffer can be located in any of these two areas. Attacker provides data
as input to the program to store in a buffer. This data is actually the instruction with
the help of which attackers try to use victim program’s buffer to store the malicious
code of his/her choice.
Attacker injects this executable malicious code into the buffer and overwrites the
return address with the address of this malicious code as shown in Fig. 4b. This
return address can be chosen by hit-and-trial method.
306 M. Khurana et al.
Fig. 4 a Malicious code can be inserted in buffer. b Return address jumps to malicious code
1.3.1 There are some difficulties with buffer overflow attack, they are [5]
(a) Attacker may not know the exact location of malicious code injected.
(b) Attacker may not know the exact location of the return address with malicious
code starting address.
1.3.2 These difficulties can be overcome by different methods:
(a) First problem can be solved by injecting No Operation (NOP) before malicious
code.
(b) Second problem can be resolved by inserting the return address repeatedly. This
may overwrite the actual return address with attacker’s return address and will
make pointer jump to any NOP address which in turn may point to next NOP
and after last NOP malicious code will be executed (Fig. 5).
There are some ways to protect the software from buffer overflow:
(a) Brute force method—to write completely the correct code but to write an
error-free code is not achievable. One of the ways to achieve near to error-free
code is to introduce buffer overflow intentionally to search for vulnerable
components in the program. The other way is to use debugging and analysis
tools to find buffer overflow vulnerabilities. This method does not eliminate all
the vulnerabilities but reduces them [6].
(b) Do not allow the code to execute on stack; stack is made non-executable by
using No execute bit or NX bit (supported by some hardware); memory can be
flagged so that code cannot be executed in a specified location.
(c) Safe program language—Java and C# have boundary check at runtime. They
automatically check the arrays out of bound. These languages do not allow
memory locations to be accessed which are out of boundary but have perfor-
mance penalty for checking, due to which developer chooses C language. So in
that case do not use unsafe function, use its safe alternative. Use safe functions
such as fgets, strncpy, strncat, and snprintf instead of C unsafe functions gets,
strcpy, strcat, sprintf, scanf, etc. [7].
(d) Runtime Stack Checking—Runtime stack checking can be introduced by
pushing special value on the stack after return address. When return address is
popped off stack, the special value can be used to verify that return address has
not changed and in order to overwrite the return address, this special value also
needs to be overwritten.
3 SQL Injection
SQL injection is one of the most famous attacks used in hacking. Every web
application has its data stored in any database. These databases contain some
sensitive and confidential data. Web applications accept the data from the users.
This data is retrieved from the database through SQL queries. To insert, retrieve,
update, and delete the data from database, SQL language is used. Using SQL
injection attacker can have unauthorized access to the system. For example, there is
any website let say Gmail that provides user an interface to enter his email id and
password. The email id and password form the part of the internal SQL query. User
enters his credentials, then these credentials are matched with the data stored in the
database. So if the hacker gets the access to that database he can easily get your
credentials and thus can attack your account like sending fake mails, deleting
important data, or extracting private information from database. Thus, SQL injec-
tion is defined as a mechanism that allows hacker to inject SQL commands to allow
them to gain access to the data held within your database. SQL injection uses the
308 M. Khurana et al.
concept of duality of data and command to get information from database. In SQL
injection, the hacker type SQL keyword to modify the structure of SQL query was
developed by web programmer, and trick the SQL interpreter to execute uninten-
tional orders. The SQL query is modified in such a way that the interpreter is unable
to differentiate between the actual command and hackers input [8]. The interpreter
is tricked to execute such unintended commands. For example, when we search any
website, we write URL:
Original query: https://88keystoeuphoria.com/video.php?id=’32’
Injected query: https://88keystoeuphoria.com/video.php?id=’32’’
This is translated into query—Select * from TABLE where id=’32’’
The hacker has intentionally modified the query and inserted an extra apostrophe
after 32. It is syntactically incorrect, so our database will throw an error message
that infers the information about table like table name. Therefore, he is able to
extract information using wrong query.
For SQL injection, first the hacker/attacker identifies whether a website is vulner-
able or not. There are various tools to check vulnerability of website [9]
(a) Acunetix—It automatically checks the given web application for SQL injection
and other vulnerabilities.
(b) Nesus—Nessus is the best unix vulnerability testing tool. It also runs on
windows. Key features of this software include remote and local file security
checks client/server architecture with a GTK graphical interface etc.
(c) Retina—It is an another vulnerability testing tool. It scans all hosts on a net-
work and reports on any vulnerability found.
(d) Metasploit framework—It is an open-source penetration testing software tool
with the world’s largest database of public and tested exploits.
For example, let say we have a website say keystoeuphoria.com. Now we are
going to exploit it using Kali Linux. Before starting, read the following disclaimer:
You may face legal action if you do not have the permission from the admin-
istrator of the website that you are testing for SQL injection. They can track your IP
address. So it is advisable to try it only if access privilege is provided.
SQL injection includes four main steps [10]:
Step 1: Enumerate the database
Open the terminal window on Kali Linux and write the following command
Sqlmap—u “https://keystoeuphoria.com/video.php?id=32”—dbs;
Result—The command checks whether the typed URL website is vulnerable or
not, and if it is vulnerable it will show you the list of various databases that exist
over that website (Fig. 6).
Buffer Overflow and SQL Injection … 309
4 Conclusion
Buffer overflow and SQL injection are still biggest security problems in software
and web applications, respectively, that will exist in future for long time due to
large amount of legacy code. This paper explains buffer overflow attack vulnera-
bilities and the preventives measures that can be taken to protect it from the
attackers. This paper demonstrates method with example for testing web applica-
tions for SQL injection vulnerabilities that attackers use to compromise a web
application. These SQL queries can be tried on real-time application under
administrative control.
Refrences
Author Biographies
Mehak Khurana is currently working as assistant professor in The NorthCap University in CSE
and IT and has around 6 years of experience. She completed her M.Tech from USIT, GGSIPU in
2011 and B.Tech from GTBIT, GGSIPU in 2009. Currently she is also pursuing Ph.D in the field
of Information Security and Cryptography at NCU. Her key areas of interest are Cyber Security,
Ethical Hacking and Cryptography. She has contributed research papers in various national and
international journals and conferences. She is lifetime member of Cryptology Research Society of
India (CRSI).
Ruby Yadav has worked as Research Associate in The NorthCap University in CSE & IT dept.
She has published papers in reputed international conferences and journals. She has completed
M.Tech and B.Tech from MDU. She is lifetime member of Cryptology Research Society of India
(CRSI).
Meena Kumari has worked as a professor, Dept of CSE & IT at The NorthCap University. She
has also worked as Scientist ‘G’ at DRDO (Defence Research & Development Organization) and
has 37 years of research experience in cryptology.
Prime Numbers: Foundation
of Cryptography
Abstract Prime number plays a very important role in cryptography. There are
various types of prime numbers and consists various properties. This paper gives
the detail description of the importance of prime numbers in cryptography and
algorithms which generates large/strong prime numbers. This paper also focuses on
algorithms which find prime factors and tests whether the entered number is prime
number or not.
1 Introduction
Exchange of information or data plays a very vital role nowadays. There are various
ways through which this data is exchanged. Today’s most common way is to
communicate through some electronic medium for exam Internet. We perform
many important tasks through the internet such as online shopping, online banking,
personal data share, etc. So it is very important to make this communication very
secure so that an attacker will not be able to get access to the data. Currently, there
are various security measures to make this communication secure one of the method
is to use Cryptography [1–4]. Cryptography focuses on the concept that “security
can be achieved by hiding the data or converting it into some unreadable form,” so
cryptography is the study of mathematical science which is used to convert the data
in some incomprehensible form which gives security to the data [1–3], i.e., cryp-
tography is the art of secret writing [4]. Secret writing is achieved by applying the
key to the original data which converts original data into unreadable data (called as
Encryption) and unreadable data to original data (called as Decryption) [1–4]. This
task is achieved by applying key at sender and key at receiver for encryption and
decryption. Two types of keys are mostly used in cryptography, symmetric key and
asymmetric key [3, 4]. The intensity of the security will completely rely on the type
of key is used. An asymmetric key is much stronger then symmetric key as in
asymmetric key two different keys are used one for encryption (encryption key is
publically declared) and one for decryption (Decryption key is private only known
to receiver) whereas in symmetric key, the same key is used to encryption and for
decryption [1–4]. An asymmetric key is also called a public-key cryptosystem
[3–5].
There are various algorithms which are used to provide security to the data.
Basically, all the concepts of cryptography based on the modular arithmetic con-
cepts, Number systems, Groups rings, Fields, etc [4, 5]. This paper focuses on the
concepts of the number system and in which prime number which plays a vital role
in Cryptography. If we consider about asymmetric key then the calculation of key
completely depends on the prime number and its factors [3–5].
2 Prime Numbers
The numbers which are divisible by itself or by 1 are called as prime numbers and
other numbers are called as composite numbers. Examples: 2, 3, 5, 7, 11, 13, 17,
19, etc., are prime numbers which are divisible by only one or by itself and rest of
the numbers such as, 2, 4, 9, 10, 12, 14, etc., are composite numbers [4, 6–10]. The
securities of cryptographic algorithms are depending on prime numbers and its
length. There are various type of prime numbers such as Balance prime, Circular
Prime, Long Prime, Mersenne Prime, Minimal Prime, Strong Prime, Palindromic
Prime, Permutable Prime, Twin Prime, Unique Prime, Wilson Prime, Regular
Prime, Integer Sequence Prime, Higgs Prime, etc. All these types of prime numbers
have different properties and it is used in cryptography depending on its properties.
The main type of prime numbers which plays a vital role in cryptography are strong
prime numbers. A strong prime is a prime number with certain special properties.
A number p is a strong prime number if it satisfies following conditions [2–4]:
• p is large prime number
• p − 1 must have large prime number factor, say a1 q1, where p = a1 * q1 + 1
• q1 must have large prime factors say a2 q2, where q1 = a2 * q2 + 1
• p + 1 must have large prime factors say a3 q3, where p = a3 * q3 − 1.
Prime Numbers: Foundation of Cryptography 317
4 Primality Testing
There are various tests which will give us result that whether the entered number is
a prime number or not. Methods such as Fermat little’s theorem, Miller Rabin,
Solovay Strassan [2–6]. An old method of primality checking on the given number
is trial and error method, where number “n” will be divided by all possible m from 2
to n, if n gets divided by m then the number is not prime number else number is a
prime number [2–5].
Example:
n = 13 then m = 2, 3, …12
n mod m =? 13 mod 7=6
13 mod 2 =1 13 mod 8=5
13 mod 3 =1 13 mod 9=4
13 mod 4 =1 13 mod 10 = 3
13 mod 5 =3 13 mod 11 = 2
13 mod 6 =1 13 mod 12 = 1
b. Bimultiplicativity:
m1m2 m1 m2 m m m
¼ or ¼
n n n n1n2 n1 n2
c. Invariance:
m m mod n
¼
n n
e. Special Values:
2 ð n2 1Þ=8 1 0
¼ 1 ; ¼ 1; ¼0
n n n
Step 3:
y ¼ aðn1Þ=2 mod n
If z y mod n
Then
Write (“Entered number n is prime”)
Else
Write (“Entered number n is composite”)
Example:
Say n = 367
21 73 7 3
x¼ ¼ ¼ ¼ ð1Þð1Þ ¼ 1
367 367 367 367
x ¼ 1
Step 3:
n1
y¼a 2 mod n
3671
y ¼ 21 mod 367
2
ð3Þ
y ¼ 366 ðWhich is equal to 1 mod 367 ¼ 366Þ; so
y ¼ 1
x y mod n
Example:
n = 131
322 S. Sarnaik and B. Ansari
If we consider the example of RSA algorithm we can say that security of RSA
completely depends on the two prime numbers, but is very difficult to find such
strong prime numbers because if a prime number is week then the decryption key
will easily break [11]. Similarly, there are various algorithms in cryptography which
uses the prime number in the process of key generation. To avoid this difficult to
find large or strong prime number, there are various prime number generation
algorithms which gives a strong/large prime number as output [4, 12]. Algorithms
such as a naive incremental generator, Random search for a prime Product of
Primes, Modular search method, Williams–Schmidt algorithm for finding strong
primes, Gordon’s algorithm for finding strong primes, etc [3–10, 12]. If we discuss
Gordon’s algorithm for finding strong primes then the following algorithm and its
output shows how it produces large and strong prime number from two small prime
numbers [12].
Step 1: Pick an integer j0. Calculate and pick the first prime number in
the sequence of 2 * j * r + 1, Where, j = j0, j0 + 1, j0 + 2 …
Denote this prime by s = 2 * j * r + 1
Step 2: Calculate l0 = 2(q * s − 2 mod s) q − 1.
Step 3: Pick an integer k0. Calculate and discover the first prime number
in the sequence l0 + 2k * s * q,
Where k = k0, k0 + 1, k0 + 2 . . .
Symbolize this prime by p = l0 + 2 * k * s * q.
Step 4: Write (p).
Prime Numbers: Foundation of Cryptography 323
Example:
It is easy to find Prime factors of a small number, such as 35 = 7 * 5, but the same
task becomes difficult if we try on very large numbers. In public-key cryptography,
it is very important to get prime numbers through factorization from a large
composite number [4, 5, 8, 13–17]. Various algorithms are there which performs
the task of finding prime factors of a large composite number, such as Number Field
sieve, Quadratic sieve algorithm, Pollard P-1 algorithm, Pollard’s rho algorithms,
etc [3–5, 8, 16, 17]. We can get the original odd composite number by multiplying
Prime factors with each other. Consider the following example Where, 8633 is an
odd composite number and 89, 97 are two prime factors, By multiplying these two
factors, we can get the original odd composite number, 89 * 97 = 8633.
Example: 8633 = 89 * 97, Here n 8633, p 89 and q 97
i. Pollard’s rho algorithm:
Integer factorization algorithms can be differentiated in two terms,
Special-purpose algorithm and general purpose algorithm, Pollard’s rho algorithm
is an example of special-purpose factoring algorithm, which is used to find small
prime factors of a composite integer. It is basically useful to find nontrivial factors
[3, 4].
Here, p and q are the smallest prime factors of n. This algorithm uses polynomial
function f with integer coefficient, i.e., f(x) = x2 + c, Where c can be any value from
1 but not c 6¼ 0, −2 [3–5, 13–17].
Example:
Example:
7 Conclusion
Book References
1. Menezes B. Network security and cryptography: Cengage Learning, India, 2010, 432
2. Bose R. Information theory, coding and cryptography 2008, Tata Mc Graw hill
3. Menezes AJ, van Oorschot PC, Vanstone SA (2001) Handbook of applied cryptography,
CRC Press, London, Oct 1996, 816
4. Stinson DR (2006) Cryptography: theory and practice, 3rd edn. CRC Press, London
Journal References
5. Rivest R, Shamir A, Adleman L (1978) A method for obtaining digital signature and
publickey cryptosystem communications. ACM 21:120–126
6. Crandall R, Pomerance C (2001) Prime numbers, a computational perspective. Springer,
New York
7. Joye M, Paillier P, Vaudenay S (2000) Efficient generation of prime numbers?,
Springer-Verlag, 1965:34–354
8. Rivest RL, Silvermany RD. Are strong primes needed for RSA?
9. Agrawal M, Kayal N, Saxena N. Primes is in p
10. Wagsta SS Jr (2014) Is there a shortage of primes for cryptography?, 2(IX), Sep 2014,
IJARET
11. Sarnaik S, Gadekar D, Gaikwad U. An overview to integer factorization and RSA in
cryptography
12. Saouter Y. A (1995) new method for the generation of strong prime numbers, RR-2657,
INRIA
326 S. Sarnaik and B. Ansari
13. Galbraith SD (2012) Towards a rigorous analysis of Pollard Rho. Mathematics of public key
cryptography. Cambridge University Press, Cambridge, pp 272–273, ISBN 9781107013926
14. Yan Y (2008) Integer factorization attacks. Cryptanalytic attacks on RSA, Springer-Verlag,
US, 255
15. Abubakar A, Jabaka S, Tijjani BI (2014) Cryptanalytic attacks on Rivest, Shamir, and
Adleman (RSA) cryptosystem: issues and challenges, JATIT, Mar 2014, 61(1):37–43
16. Hawana B (2013) An overview and cryptographic challenges of RSA. IJERMT
17. Chalurkar SN, Khochare N, Meshram BB (2011) Survey on modular attack on RSA
algorithm, IJCEM, Vol 14, Oct 2011, 106–110
Steganography: A Survey
Shilpa Pund-Dange
Abstract Due to the popularity of Internet and communication media, data secu-
rity is one of the most focused areas of research. Cryptography and steganography
are two important disciplines in the area of security. Image steganography is one of
the techniques used to hide data inside an innocent image so that it is not visible to
an eavesdropper while communication. This is a very useful technique when we
transmit data from sender side to the receiver side. Many steganographic methods
are suggested by the researchers struggling for good embedding capabilities and
imperceptibility. This paper describes a survey on various image steganography
techniques in spatial domains with their strengths and weaknesses.
1 Introduction
In recent years, the security and confidentiality of sensitive data have become very
important due to the fast growth of internet and communication technologies.
Therefore, how to protect this secret data from the unauthenticated user during
transmission become an important issue nowadays. The well-known techniques
which are used to provide security are Cryptography and Steganography.
Cryptography is often used to protect information secrecy making message
scramble so that it cannot be understood. Steganography means hiding information
in digital media such that no one should be aware of the existence of information.
S. Pund-Dange (&)
Department of Computer Science, Modern College, Shivajinagar,
Pune 411005, India
e-mail: [email protected]
2 Steganography Techniques
There are three different ways to hide a digital message in a digital cover.
Injection: In injection method, the secret message is directly embedded in the host
medium. This increases the file size and hence sometimes the changes can be easily
detectable.
Substitution: In this method, the actual data is substituted with the secret data. This
creates a little change in the size of the cover object. However, the quality of the
cover object can degrade depending on the type of cover object and the amount of
data embedded.
Generation of New Files: In this method, a cover is specially generated for the
purpose of concealing a secret message [1].
Steganography
JPEG Patchwork
There are many versions of spatial steganography. In all these methods, some of the
bits in the image pixel values are directly changed for data hiding. Least significant
bit (LSB) steganography is one of the simple techniques that hide a secret message
in the least significant bit of the pixel value. As the only LSBs are changed, there is
no distortion in the image and the change is not perceptible to the human eye. Some
spatial domain methods are listed below:
1. Least significant bit (LSB)
2. Edges-based data embedding method (EBE)
3. Pixel value differencing method (PVD)
4. Pixel intensity-based method
5. BPCS steganography
6. Mapping pixel to hidden data method
7. Labeling or connectivity method
8. Random pixel embedding method (RPE)
9. Texture-based method
10. Histogram shifting methods [2, 3].
Generally, image steganography is categorized into the following aspects.
330 S. Pund-Dange
3 Literature Review
As the focus is on Spatial Domain Method, some of the methods are explained
below:
In LSB method, a digital media like audio, video, or image, there is a large
amount of space which we can use for steganography. Digital data consists of bytes.
Each byte contains 8 bits. These 8 bits makes a color of a pixel. The MSB plays
important role in the different shades of color. The LSBs have less impact on color.
So if we make a change in the Least Significant Bit, it changes the value by +1 or
−1 which is not perceptible to the human eye. So by taking the advantage of human
perception, LSB steganography works.
For example, following is the bit representation of the digital cover
We want to embed a message which is a secret code, suppose 207. The binary
representation of 207 is 11001111.
Now by using the LSB method, the message is embedded as follows:
To embed 8 bit of data we need to change only five bits. This change will create a
very small or no noticeable difference in the cover image. Hence, if the digital cover
is in several kilobytes or megabytes then we can embed huge amount of data within.
In Pixel Value Differencing (PVD) method [4], the cover images having the
maximum intensity value 256. Two neighboring pixels p1 and p2 are read and the
difference value d in between them is computed. The reading of two neighboring
pixels of the cover image is carried out through each of the rows of the image in a
zigzag manner. Let the gray values of two adjacent pixels are g1 and g2 then
d = g2 − g1. Take the absolute value of d which may be in the range from 0 to 255.
Steganography: A Survey 331
All values of d are in the range 0–255 say Di. The width of Di is calculated which is
always taken as a power of 2. The secret message is embedded according to the
range of Di and is replace with another difference value d. The difference ‘n’
between the old and the new difference values is calculated. And accordingly, the
new pixel values are calculated as gi-ceiling (n/2) and gi+1 +floor (n/2). The mes-
sage bits are extracted by calculating the difference between the new difference
value and the lower bound of the range block.
In ELSB [5], all the edge pixels in an image are used. Here, the masked image is
created by masking two LSBs in the cover image. Then, find out the edge pixels by
using the Canny Edge detection method and then hide the secret message in the
LSBs of the edge pixel. In this way, the stego-image is formed.
At the receiver, the stego-image is again masked as the same. Then by using
canny edge detector, the edge pixels are identified. We will get same edge pixels.
The secret message is extracted from the two LSB bits of the identified edge pixels.
BPCS [6, 7] by Eason, overcomes the limitations of LSB technique. As com-
pared to the above techniques, BPCS Steganography has very large embedding
capacity. In LSB technique, data is embedded in LSBs. But in BPCS technique data
can be embedded in planes in the complex region.
An image consisting of n-bit pixels. Convert all pixel intensity values in binary.
If n = 8, then every pixel is 8 bits. Then decomposed the image horizontally into
8-bit planes. Therefore, img = [PL7 PL6 PL5 PL4 PL3 PL2 PL1 PL0] where PL7 is
the Most Significant Bitplane and PL0 is the Least Significant Bitplane. For each bit
plane, the complexity of the image is determined. Accordingly, an image is seg-
mented into the informative region and noise-like region. An informative region is
having a simple pattern while noise-like region having a complex pattern. Here,
data embedding takes place in the noise-like region. Thus, BPCS steganography is
not perceptible to the human vision system.
Pixel Intensity Based method [8], uses 24-bit RGB image. So, Three pixels
(Red = 255, Green = 255 and Blue = 0) are generating Yellow color. If we change
(Blue = 16) still it generates Yellow color. If both yellow colors are comparing,
both have almost the same visibility. So, the idea is that if we change the lower
intensity pixel value it has less visual degradation quality effects. So here we can
use 4 LSBs for data embedding.
4 Observations
LSB method
• LSB work well with grayscale as well as color image.
• LSB method is easy for implementation.
• Changes in the image are not perceptible to the human eye.
• But once notice suspicious, easy to crack the message.
332 S. Pund-Dange
This paper describes different techniques and types of steganography. Also, it gives
a survey on different steganographic methods for the image in spatial domain with
some observations and limitations. The next plan is to develop a hybrid stegano-
graphic method using the existing methods or a new one which satisfies
steganography aspects mentioned above.
Steganography: A Survey 333
References
Author Biography
Keywords Threat model Vulnerability assessment Malicious node
Asset- and attacker-centric model Black hole Wormhole
1 Introduction
Threat to a network is any potentially malicious occurrence that can disrupt the
operation, functioning, integrity, or availability of the network or system [1, 7, 8].
There are three general approaches to threat identification: (a) Attacker-Centric
approach views it from an attacker’s point of view–their goals, motivations, and
how they might achieve them; (b) Design-centric approach starts from the design
of the system looking for types of attacks against each element (c) Asset-Centric
approach starts from assets entrusted to a system [9].
Ad hoc networks require customized threat model in contrast to threat model of
traditional networks. Therefore a hybrid threat identification methodology was
developed which combines the ‘Asset centric and Attacker centric’ approaches that
take into account the critical assets present in the network and the capabilities of the
adversary required to compromise these assets.
Comprehensive Methodology for Threat Identification … 337
Ad hoc network deployment and functioning require a set of resources that are
shared among nodes. These can be described as assets. Although the network has a
number of assets, only those critical assets have been chosen that are fully owned
by the ad hoc layer [10] and are discussed below.
(a) Algorithm Processing: These are resources deployed in a node for calculating,
maintaining, and processing ad hoc networking.
(b) Algorithm Storage: The space required to store algorithms for the node that
are loaded on booting time or on request.
(c) Network and User Information: This refers to information shared between
nodes to aid in routing and contains information such as node location, power
availability, node speed, direction, radio profiles, user profiles, etc. This also
includes the information about the routing tables stored on a node.
(d) Network Topology and Node Roles: This refers to information about the
topology of a network, its behavior, and function of individual nodes and their
routing loads.
(e) Payload Messages: These are messages containing the control information and
user data which is carried on behalf of an application.
(f) Routing Messages: These are route discovery, update, and reporting messages
that are critical for an ad hoc network to successfully maintain connectivity and
routing capabilities.
As the physical layer is wireless in ad hoc network, the adversary can exploit it in
many ways and disrupt the network functions [6–8]. The adversary’s capabilities
can be characterized as
(a) Passive and Active: A passive adversary listens and records all the messages
including the routing updates in an unauthorized way. This information can be
used for traffic analysis. An active adversary prefers to interfere in some way,
e.g., by modulating packet, forwarding, injecting, replaying packets, etc.
(b) External and Internal: An external adversary carries out attacks by nodes that
do not belong to the network. It causes congestion, sends false routing infor-
mation or unavailability of services. An internal adversary mounts attacks as
compromised nodes that are part of the network.
(c) Static and Mobile: A static adversary has the capability to set the corrupted
nodes only once. A mobile adversary has the capability to change the set of
corrupted nodes from period to period.
338 R. Tyagi et al.
In the proposed solution, threats have been categorized under following three
categories:
(i) Host-based threats, (ii) Network-based threats, (iii) Application-based threats.
And for each category threat identification was carried out taking into account
the targeted assets and required attacker’s capabilities [11]. The security parameter
breach by the adversary has also been considered.
The first step of the methodology was to classify critical assets under these three
categories. Table 1 categorizes the assets of ad hoc network.
The developed threat identification method was applied to each category. It gives
a list of threats and possible attacks under individual threat, the impact of the threat
on the critical assets, attributes of the adversary’s capability required for specific
threat and security parameter breached by the adversary. Threats against
network-based assets are described in Table 2. The major threats affecting the
network are Eavesdropping, Manipulation of data packets, Routing protocols threat,
Misdirecting traffic, DoS, and Masquerading.
Threats against Host-based assets are described in Table 3. The major threats
affecting the hosts are DoS, Modification, and Information leakage. Threats against
Application-based assets are described in Table 4. The major threats affecting the
applications are Manipulation of application protocols and services, DoS and
Repudiation of Services.
Comprehensive Methodology for Threat Identification … 339
Table 2 (continued)
Threat’s name Critical assets affected and Attributes of adversary’s Security
and possible threat’s description capability required parameter
attacks breached
Routing Routing messages: Aim to Active, external/internal, Availability
protocol disrupting the operation of static/mobile, deployment integrity
threats the network by modifying capability
(i) Routing routing information
table (RT) (i) Adversary advertises
overflow routes and prevents
(ii) RT creation of new routes to
poisoning authorized node
(iii) Packet (ii) Compromised nodes
replication send fictitious routing
(iv) Route updates
cache (iii) Adversary node
poisoning replicates stale packets
(v) Rushing (iv) Poison the route cache
attack (v) Adversary rapidly
spreads routing message
in the network. In each
route discovery, acts as
one of the intermediate
node
Misdirecting Routing messages & Active, internal, static/ Integrity
traffic payload messages: Redirect mobile, computational Availability
(i) Black hole traffic to a different bounded/unbounded,
(ii) Gray hole destination Byzantine, deployment
(iii) Wormhole (i) A malicious node falsely capability
(iv) Byzantine advertises good path then
intercepts and discards all
packets
(ii) A malicious node has the
ability to forward some
routing packets and
discard others
(iii) A tunnel is generated
between two colluding
attackers and redirects
the traffic through this
(iv) Set of compromising
nodes works in
collusion to create
routing loops,
nonoptimal routing
paths and dropping
packets
Denial of Algorithm processing and Active, external/internal, Availability
service network topology and node static/mobile, deployment
(i) Jamming roles: Adversary attempts to capability
(ii) Flooding
(continued)
Comprehensive Methodology for Threat Identification … 341
Table 2 (continued)
Threat’s name Critical assets affected and Attributes of adversary’s Security
and possible threat’s description capability required parameter
attacks breached
(iii) Distributed prevent legitimate users to
DoS access network services
(i) Adversary transmits high
power signal of same
frequency at which node
is receiving signal
(ii) Adversary floods the
network with false
routing messages
(iii) Several adversaries that
are distributed
throughout the network
collude and prevent
legitimate user from
accessing the services
Masquerade Network topology and Active, external/internal, Authentication
(i) Spoofing node roles: Adversary static/mobile, deployment Confidentiality
(IP, MAC) assumes the identity and capability Integrity
(ii) Sybil privileges of an authorized Anonymity
(iii) Session node
hijacking (i) Adversary modifies
address information in
packets and adopts an
authenticated identity in
the network
(ii) Adversary represents
multiple false identities
(iii) Adversary takes control
over a session between
nodes
Table 3 (continued)
Threat’s name and Critical assets affected Attributes of Security
possible attacks and threat’s description adversary’s parameter
capability required breached
Modification (i) Algorithm storage: Active, external/ Integrity
Data corruption, node Algorithms may be read internal, deployment Authenticity
malfunction, node or altered at the node’s capability Confidentiality
replication storage Anonymity
(ii) Network and user
information: nodes or
user specific
information might be
modified
Information leakage (i) Network and user Passive, external, Confidentiality
Traffic monitoring and information: Node or static, computational Anonymity
analysis user specific bounded/
information might be unbounded,
readable deployment
(ii) Payload messages: capability
Unauthorized capture of
transmitted data might
be possible
Table 4 (continued)
Threat’s name and Critical assets affected and Attributes of Security
possible attacks threat’s description adversary’s parameter
capability breached
required
(ii) Environmental (i) In this approach, collusion
(iii) Service collusion by set of nodes to isolate a
threat target
(ii) In distributing sensing
services, environmental
corruption of the sensed
information may prompt
the derivation of a
misleading inference
(iii) Set of nodes
compromised and gives
misleading information
to services (sensing)
Denial of service Network topology and node Active, external/ Availability
(i) Resilience thrashing roles, algorithm processing, internal, mobile,
(ii) Resources and storage: Services deployment
consumption (buffer running at node may be capability
limit, computational subjected to DoS attack
power) (i) In resilience services,
MANETs should be
dynamically
reconfigurable and
services may be
dynamically relocated. By
clever manipulation,
adversary causes a system
to repeatedly reconfigure
(ii) Adversary needs to send
appropriately timed
requests to deny service
availability to legitimate
users. It also sends
service requests that are
highly computationally
intensive
Repudiation of a Payload messages: Denial of Active, external/ Non-repudiation
service service by an adversary node internal
involved in communication
The possibility of a node becoming malicious is very high and is the second major
source of vulnerability in the ad hoc networks. A node can become malicious when
it breaches any of the security principles and start behaving in one or more of the
following ways:
(i) A node drops the packet fully or partly.
(ii) A node wastes the battery, storage, and bandwidth by performing unneces-
sary operations.
346 R. Tyagi et al.
(iii) A node becomes a part of the network without authentication and starts
disrupting the normal behavior of the network.
(iv) A node starts injecting stale packets and creates confusion in the network.
These vulnerabilities lead to Wormhole, Byzantine, Resource consumption,
Impersonation, Session Hijacking, and DoS attacks.
Crypto algorithm forms the core of security in ad hoc networks. The security
parameters––authentication, confidentiality, and integrity of data should be taken
care during design of the network. The strength of the cryptographic algorithms
used should be high and match with the level of secrecy required in communication.
Inappropriate and weak crypto algorithms can be exploited by an adversary to
compromise a node or data and routing protocol communication and become a
source of vulnerability. These vulnerabilities lead to unauthorized access to the
network, system, and data and cause message tampering, stealing information, and
DoS attacks.
4 Conclusion
References
1. Murthy CSR, Manoj BS (2004) Ad hoc wireless networks architectures and protocols.
Published byPearson Education (ISBN 81-297-0945-7)
2. Sarkar SK, Basavaraju TG, Puttamadappa C (2008) Ad hoc mobile wireless networks
principles, protocols, and applications, 22. Aurebach Publications, Taylor & Francis Group
Comprehensive Methodology for Threat Identification … 347
3. Sen S, Clark JA, Tapiador JE (2010) Security threats in mobile ad hoc networks. J Dept
Comput Sci, Univ. of York, UK, 32
4. Goyal P, Parmar V, Rishi R (2011) MANET: vulnerabilities, challenges, attacks, application.
JCEM, 11:32–37
5. Sangwan S, Jangra A, Goel N (2011) Vulnerabilities and solutions: mobile ad hoc networks
for optimal routing and security. J GRCS 2(5):8–12
6. Yau P-W, Mitchell CJ (2003) Security vulnerabilities in ad hoc networks: research
programme of the virtual centre of excellence in mobile & personal communication. In:
Proceeding of the 7th ISCTA, 99–104
7. Spiewak D, Engel T, Fusenig V (2006) Towards a threat model for mobile ad hoc networks.
In: Proceeding of the 5th international conference on information security and privacy, 35–40
8. Clark JA, Murdoch J, McDermid JA, Sen S, Chivers HR, Worthington O, Rohatgi P (2007)
Threat modelling for mobile ad hoc and sensor networks. In: ITA conference
9. Information Security Provider and Research Centre (2011) Threat modelling. www.
Praetorian.com
10. Martin A (2006) A platform independent risk analysis for mobile ad hoc networks. In: Boston
university conference on information assurance and cyber security
11. Chidambaram V (2004) Threat modelling in enterprise architecture integration. SETLabs
briefings. EABC 2(4)
12. Saini R, Khari, M (2011) Defining malicious behavior of a node and its defensive methods in
ad hoc networks. JCA, 20(4)
13. Chayal D, Rathore VS (2011) Assessment of security in mobile ad hoc networks (MANET).
J GlobResComp Sci, 2(6)
14. Kayarkar H (2012) A survey on security issues in ad hoc routing protocols and their
mitigation technequies. Int. J. Adv Netw Appl 03(05):1338–1351
15. Sen S, Clark JA (2007) Intrusion detection in mobile ad hoc networks: In guide to wireless ad
hoc network. InGuide to Wireless Ad Hoc Networks, 53:427–454. Springer Publication,
London
16. Marti S, Giuli TJ (2000) Mitigating routing misbehaviour in mobile ad hoc networks. In:
Proceedings of the 6th ACM international conference on mobile computing and networking
(MobiCom), pp 255–265
Hardware Trojans: An Austere
Menace Ahead
Keywords Integrated circuits Hardware Trojans Threats Networking threats
IC fabrication Backdoors System on chip Trojan side channel
1 Introduction
We all are purview to the City of Troy story wherein few hundred years back,
Greek soldiers undertook many attempts but unsuccessfully to capture the city of
Troy. Eventually, they departed, leaving behind a large wooden horse, ostensibly as
a gift. The citizens of Troy were too happy to accept the wooden horse but as it had
to come about; a group of Greek soldiers came out of the horse late night handily
and opened the gates for their paisanos, who easily dismissed the quiescent city.
Come to present, Trojan [1] as a term today is synonymous more with the IT
Security incidents that have seen a phenomenal increase over a decade. For over a
decade now, the IT Security domain loyalists have dedicated their energies,
resources, domain knowledge, brainstorming sessions and investments into ensur-
ing that the security is ensured for the user. And so the market today got an overplus
of options too, viz., antivirus solutions, Firewalls, Internet Security Editions,
UTMs, and the list goes on. These may be different technically in operating but
there is one common thing in all these options that they all have a mechanism to
detect the threats which are all software based. They have no way, no mechanism to
thwart, or even think to detect a threat which is embedded deep inside the IC
hardware. A threat is so obliterated to be seen, so unthinkable that for the panic
struck solution providers it is like where to start from? How to do? What to do?
Hardware Trojan came into being primarily imputed to outsourcing the fabrication
and design to third parties attributed to the huge scales of requirements and
economies involved. Now, this small modification can be in place anywhere of a
corporate house infrastructure, household chores appliances, or even military and
defense COTS equipment.
The austere consequences of Hardware Trojans are well left to the imagination of
what holds on to be excluded today in the increasing scenario where dependence on
IC and SoC is only increasing. The key heads affected and vulnerable to such
attacks may include Logistics Systems and Support domain, viz., Transport
infrastructure, Traffic Control, Metro/Rail monitoring and control, Civil critical
Hardware Trojans: An Austere … 351
As brought out from the above about HT, the comparison between severity
viz-a-viz Software Trojans allows HT to take leaps out-front lead. It empathizes that
the software threats that exist with us over decades now are yet to get a stable and
an assured solution by any means and this HT threat has just arrived in the fora.
A mini comparison [3] between the two is bought out in the figure below:
A hardware Trojan to operate needs ground and power supply which can be low or
high depending on the design it is based on. A Trojan that requires a low-end power
supply will have low chances of being detected whereas a Trojan requiring higher
power supply would invariably be at a larger chance of detection by a sensor if
placed. Hardware Trojans have a range of classification based on various charac-
teristics and modes they work in. The classification keeps on evolving as more
newfangled approaches and dimensions of attacks are detected. A form of classi-
fication based on the activation mechanism of triggering the attack that can be
Digital or Analog. Analog will typically get activated based on any analog input
type like Temperature, Pressure, time-lag etc. whereas digital will be based on some
kind of Boolean logic function [4] has classified the same in another manner as
shown in Fig. 1. The classifications as shown are to some degree perceivable by
352 A. Tiwari and C. Soni
Hardware Trojans
Type
their names and depict a very basic and broad classification of the HTs. Another
classification [5] divides Hardware Trojans into two types that include
Parasite-Based HTs and Bug-based HTs. A parasite based HT hides in the original
circuit without altering it and is not involved to lose any set and defined func-
tionalities in the circuit whereas a Bug-based HT not only alters the circuit but also
causes it to lose its set and configured functionalities. Of the two parasites HTs
become more difficult to be detected through owing to hidden nature and is actually
untraceable in specified specs as well as testing.
4 Nemesis Framework
Hardware Trojans insertion would actually gain a large mileage and suit to bestow
maximum scathe in a typical supply chain which essentially consists of unalike and
miscellanea of insertion points. Hardware Trojan-infected hardware would be more
apt for a larger organization and a huge victim base since it will allow a deeper
penetration in terms of the scale of victims.
packet timings. The activity LED light seen in general flicking gives a broad
indication of the current network traffic presently user is involved with. The [6] has
taken the RTL88111E chip for the study which deciphered that there is 160 ms
delay between the LED Light to cycle on and off and it is this 160 millisecond delay
during which there is no network activity. This timing behavior of this LED activity
is used as a trigger for the Hardware Trojan.
Further to this [7], demoed the payload execution with the ENW02A-1-BC01
Gigabit Ethernet PCI-Express card. The network hardware Trojan was shown
degrading the network services using noise injection in chips clock circuitry of the
Ethernet controller in the form of a bias voltage. The demonstration included
desynchronizing the clock of the Ethernet controller chip owing to changes in the
affected bias voltage that lead to the changes in the resonant frequency on the
external crystal. Vide [8], HT can be an elementary alteration to the original IC.
This refers to an insertion of two input AND gates wherein while the HT is inactive
the IC gives the desired output unaffected while the same gives an always zero
output irrespective of the input given. They referred this particular example as Stuck
at Zero Trojan, i.e., SAZ.
5 HT Insertions
Hardware Trojans can actually get inserted at various stages of their life cycle
typically during design and manufacturing process or maybe even retrofitted to an
existing Hardware IC.
CAD tools can be periled appositely tapping software vulnerabilities to alter RTL
[10] without the intercession and intent of the designer and once compromised, it
would be a herculean task to detect. Besides, the concept of SoC based on recy-
clable hardware is a permeating praxis in the semiconductor industry today owing
to the huge diminution in cost and time attributes involved. Sadly here, only the
supply and demand factors are being addressed, i.e., the user is only interested in
getting his functionalities right and the seller may just be involved to ensure the
same reaches the customer at the right time but the malicious untrusted third party
in the process may butt in something unknown to either that can be a reason for
chaos later.
A typical customer holds a bare manufactured IC and configures the same with the
help of a field-programmable gate array. The FPGA configuration is generally
specified using a hardware description language (HDL), similar to that used for an
application-specific integrated circuit (ASIC).
MOLES [11] aka Malicious Off-chip Leakage enabled by Side channels engineered
to leak information below the effective noise power level of the device. Vital and
critical data vide Moles can be retrieved with the assistance of spread spectrum
technology, i.e., a type of wireless communications in which the frequency of the
transmitted signal is designedly altered [12] and since the signal of the reduced
information vanishes in the noise, it becomes arduous of what data has been
transferred [13]. HT based on this actually is a novel way to designedly leak out
information.
It comes as a surprise though that such a severe threat that is currently the topic of
various forums, discussions, conferences, and research work has no such case
studies to know the impact. Whatever heard and read is all discredited and only
suspected, for e.g., Operation Orchard [15] wherein a Syrian nuclear reactor was
subjected to Israeli Airstrike, seems to have been worked out via a hidden
kill-switch function in the radar infrastructure. This functionality was then thought
to be used to disable the Syrian radars for the short duration of the attack.
Mi-grade FPGA chips, e.g., ACTEL have been a suspect of containing a
backdoor function that’s equivalent of admin debug designed into the JTAG
functionality of the subject chip IC. The subject IC Actel/Microsemi ProASIC3
chips could be used for accessing FPGA configuration using this backdoor. The
researchers confirmed that this backdoor was not present in the original firmware
loaded with the chip [16].
Vide above basic introduction we can see the kind of potent threat this brings along
and the worst part till date is no formal or assured methods exist to detect any such
threats. A typical hardware Trojan threat can actually exist in an IC as a 5–6 line
code that gets activated under predefined conditions as a set. Though at present the
severity of the threat being realized is finally forcing IT security domain to look and
seek ways to resolve. Few good but only prelim measures include the following.
Polytechnic Institute of New York University based at USA every year organizes
this competition by the name of Embedded Security Challenge (ESC) that bids two
teams in a contest wherein one team designs target system hardware and the other
team tries to identify and exploit the vulnerabilities in the target hardware [17].
356 A. Tiwari and C. Soni
Probably, as we see above, these are only too prelim measures to counter Hardware
Trojans perhaps a long way to go before a 100% trusted IC checks in before us.
Ideally any malevolent modification to any IC should be perceptible during tests
and inspections whilst pre-silicon manufacturing or post-silicon testing but that is
not an easy thing to do since the complex ICs today, with so many multiple
agencies involved at various echelons of manufacturing and design, will unlikely
have a golden model of the intact IC. Moreover, if the antagonist decides to taint
only a minuscule percentage of the complete batch of ICs being manufactured, the
complexness to detect only step-ups further. Another way out for detection involves
Nanometer physical inspection [21] which is for one very complex from point of
conduct but also is mostly not economically viable. Vide [22], the countermeasures
for HT as concentrate on three panoptic categories of countermeasures for pro-
tection against HT. These include Runtime monitoring, Design for security and
Trojan Detection approaches which attempt to arrest any kind of malicious
embedding of HT at prefabrication stages using pre-silicon test approaches or using
non destructive techniques at post-silicon manufacturing test stage. The Run Time
Monitoring approach is based on online monitoring while the circuit is in operation.
The Design for Security approach essays to make the insertion of HT at any stage
hard or facilitate detection ease during pre/post fabrication whereas the Trojan
Detection approach can be logic testing based on generating set and predefined test
Hardware Trojans: An Austere … 357
patterns and side-channel analysis for HT. Between these, Design for Security
approach may not be a very effective way to resolve the HT threat owing to the
diversity of threat classification discussed above whilst Runtime Monitoring may be
more effective since this approach can be applied for real-time monitoring.
Once the IC is fabricated and boxing concluded for use by the end user, there
remains very restricted ambit and visibility to endeavor to detect any kind of HT
presence. However, destructive reverse engineering resolves to an extent in such
cases. It involves depackaging the IC, acquiring microscopic images of each layer,
trust validating the same after rebuilding the design of the end product. This
approach uses a sample of the infected batch of ICs, thus it would be judicious to
apply this wherein infection or insertion of the HT is limited to a small percentage.
Scanning Electron Microscopy (SEM) is used to destructively delayer one chip
wherein all of the transistors and connections can be averred. Also, this approach
makes the IC under test unusable further, that’s why the name destructive came to
the fore. It takes from weeks to maybe months depending upon the complexity of
the IC under detection for giving a 100% assurance of an HT free IC.
Nondestructive methods relate to ways of detection that keeps the chip usable after
the test. Between the two, Destructive detection technique is more effective
viz-a-viz nondestructive detection technique.
One of the countermeasures against Hardware Trojans proposed by Aliyu and Bello
[23] is the use of Homomorphic Encryption and Decryption which offers brilliant
security boasts since it allows operating on data without revealing the contents
being worked at. Homomorphic encryption is a type of encryption which allows
processing of data on ciphertext and generates an encrypted result which on
decryption is valued equally to the one processed with plain text. This certainly is
an advantage plus for handling Hardware Trojans. Homomorphic encryption may
be Partial or Full where Partial Homomorphic proffers to do either multiplication or
addition on ciphertexts without unwrapping the original plaintext data while Full
Homomorphic appropriates efficacious rating of a capricious depth circuit compiled
of multiplications and additions.
358 A. Tiwari and C. Soni
9 Conclusion
IC is the basic core component of the diverse range of electronic systems being
exploited across various domains pan globe today and the growing dependence
makes it essential to ensure these ICs faithfully and sincerely perform the tasks they
are designed and fabricated for. Hardware Trojans being inserted or retrofitted at
any stage in these ICs are thus a grave threat that stands as a serious challenge today
for the IT security domain. The software industry which has been campaigning in
all gears put into ascertaining a malware/virus free application or an OS, over
decades now, is yet to reach anywhere as daily various zero days keep getting
deciphered which might be existing in an unknown quantified figure. The HT threat
actually adds to the excruciation since this is indeed indecipherable with the present
set of researches and studies did across. The future researches have a wide domain
to work on starting to explore the emerging attacks on these ICs, developing trust
validation standards for ICs being manufactured in the electronic industry and come
out with inexpugnable apt approaches to counter such threats.
References
1. Trojan at https://en.wikipedia.org/wiki/Trojan_horse_%28computing%29
2. Mitra S, Wong HSP, Wong S (2015) Stopping Hardware Trojans in their tracks
3. Bhunia S (2014) Hardware trojan attacks: threat analaysis and counter measures
4. Karri R (2010) Trustworthy hardware: Identifying and classifying Hardware Trojans. IEEE
Comput 43(10)
5. Wang X, Plusquellic J (2008) Detecting malicious inclusions in secure hardware: challenges
and solutions. In: Proceedings of the 2008 IEEE international workshop on hardware-oriented
security and trust, Washington
6. Zhang J (2014) DeTrust- defeating hardware trust verification with stealthy
implicitly-triggered Hardware Trojans
7. Shield J, Hopkins B (2015) Hardware Trojans—a systemic threat
8. Shield J, Hopkins B (2015) Hardware Trojans—a systemic threat, p 47, Para 5
9. Shield J, Hopkins B (2015) Hardware Trojans—a systemic threat, p 49, Para 5.3
10. Aliyu A, Bello A (2014) Hardware Trojan model for attack and detection techniques
11. Rad R, Plusquellic J, Tehranipoor M (2010) A sensitivity analysis of power signal methods
for detecting Hardware Trojans under real process and environmental conditions
12. Wu TF, Wong HSP, Wong S, Mitra S (2015) TPAD-hardware trojan prevention and detection
for trusted integrated circuits
13. Lin L, Burleson W (2009) MOLES—malicious off-chip leakage enabled by side-channels
14. Spread Spectrum at http://searchnetworking.techtarget.com/definition/spread-spectrum
15. Hardware Malware book By Edgar Weippl (2013) Adrian Dabrowski, Heidelinde Hobel,
p 67, para 4.2
16. King ST, Tucek J, Cozzie A, Grier C, Jiang W, Zhou Y (2008) Designing and implementing
malicious hardware. In: Proceedings of the first USENIX workshop on large-scale exploits
and emergent threats(LEET)
17. Adee S (2008) The hunt for the kill switch. IEEE Spect 45(5):34–39
Hardware Trojans: An Austere … 359
Author Biographies
Anupam Tiwari is an IT Security enthusiast and an incisive learner, holds rich experience and
qualifications in the demesne including CDAC & GFSU Certified Cyber Security Professional,
Certified Ethical Hacker with B.E and M.Tech in Computer Science from JNTU Hyderabad. He
also holds three post graduation qualifications in Information Security, ERP and Operations &
Systems and presently pursuing his research in the world of cryptocurrencies. He is a senior
member and regular contributor to articles in leading defence and engineering journals. He has
been a regular participant in National and International Seminars as a guest speaker and He is
working with the Min of Defence wherein he has variegated experience of service in IT security
implementations and conduct of Cyber Audits.
Chetan Soni is an cyber security follower. He holds vast experience and qualifications in field of
cyber security. He is B.E. in computer science and holds two PG Diploma in field of Information
security and Aeronautical engineering. He has done various certifications in field of Information
Security including CEH. He has ten years of experience in domain of Information Security. His
area of interest is implementation of firewalls and network security. He is presently working with
Min of Defence where he has implemented various cyber security measures and conducted
Information Security audits in his organization.
Cybersecurity for Supervisory Control
and Data Acquisition
Abstract SCADA stands for Supervisory Control and Data Acquisition, a com-
munication technology which collects data from distant facilities and sends control
signals to actuators. A number of factors exist that increases the risk allied with
SCADA systems. SCADA components are considered to be profoundly privileged
targets for cyberattacks through which hackers can easily hit the nation’s critical
infrastructure and economy. This paper investigates security issues of SCADA
communication protocols. In order to protect the SCADA networks, we focus on
the protocols as they were not designed with inherent security features. This paper
emphases on the security system through protocol hardening. The objective is to
modify the structure of such protocols to provide more integrity and authentication.
In the proposed structure, two algorithms are used to enhance the security and
integrity of the payload. They are discussed further in the next six sections.
1 Introduction
S. N. Shinde (&)
Department of Computer Science, C.M.C.S. College, Nashik, India
e-mail: [email protected]
R. P. Shinde
Department of Computer Science, Sinhgad College of Science, Pune, India
e-mail: [email protected]
2 Industrial Network
SCADA has advanced prerequisite with reference to reliability, uptime, and inac-
tivity as compared to the IT systems, so it is impossible to constantly apply security
measures to the information technology organization. Confidentiality, Integrity, and
Availability is the main concern for both the systems. The top priority for SCADA
systems is Availability whereas Confidentiality is for IT system. It is required to
analyze various threats and vulnerability that affects the SCDA system operation.
The topic covered in this paper is the core of automation and Industrial grids that
constitute Critical National Infrastructure. Traditionally, such systems were instal-
led standalone and did not interface with the outside world. SCADA [3] compo-
nents are considered to be privileged targets for cyberattacks through which hackers
can easily hit the nation‘s critical infrastructure and economy. Such attacks can
potentially shut down power systems, interrupt financial service and, therefore,
obstruct the essential operations of the nation. While modern communications,
computing and control systems bid remarkable openings to expand response of the
power-driven system, optimize generating station performance, and offer resilience
to failure, they also render the physical processes and systems prone to purposeful
attacks.
Protecting the SCADA systems which perform the monitoring and control
functions of utility infrastructure, such as electricity, gas, water, etc., is critical for
national security. Any vulnerability in these systems can pose serious threats and
can bring down operations of a utility. In critical applications, the appropriate
control strategy to block execution and any unknown or malicious behavior.
Security risk analysis and development of precise safety keys will help us to
understand, how this can be protected from attacks.
The protocols used in SCADA [4, 5] systems traditionally have been built with
little thought given to security. Security of SCADA system by means of protocol
hardening is a plausible solution to address such threats.
Cybersecurity for Supervisory Control … 363
It provides customized utilities like data formats, the efficient spread of acquired
data, features, and control commands [9]. This layer provides service to send/
receive messages to/from DNP3 devices. A fragment is a block of octets containing
request or response information transported between a master and an outstation.
Application layer fragment structure is of two types:
• Request fragment
• Response fragment
In request fragment, the application request header is of 2 bytes: Application
control (1 byte), Function code (1 byte).
In response fragment, the application response header is of 4 bytes: Application
control (1 byte), Function code (1 byte), and Internal Indication (2 byte).
364 S. N. Shinde and R. P. Shinde
The device in industrial networks are sometimes located in remote places and
therefore device security takes a high priority in design and deployment of these
devices.
The intense growth of liabilities has become one of the key challenges for security
personnel, who not only need to consider the increasing amount of attacks, but also
identify how these attacks could be combined in complex ways. Clearly, a
methodology must exist, but it is not significant. For instance, a company may
identify their competitor using industrial spies against them, but the company may
judge this as nonexistent for specific parts of their infrastructure. For example, an
outlying control station operating with 10-year-old technology is less prone to
threats. The severity of the threat must be determined by allocating resources
properly.
Few business experts treat SCADA as usually customized for particular sector’s
application, which requires a good precise knowledge of a certain system and
industry to attack it. Furthermore, specialized knowledge requirement will reduce
the number of attackers, concluding why SCADA attacks are diverse from the
attacks on other computer networks. Although SCADA attacks are infrequently
effective, the oppositions showcase its curiosity.
There are a number of ways using which we can perform a security analysis of a
given network based on protocol application and topology.
Fault Tree Analysis (FTA) is a tool used for security and reliable evaluation for
demonstrating the failure paths in a system. It does system-level risk evaluations
using a tree structure. It is almost 50 years old and is extensively utilized. The
failure in the system is exhibited in a visual fault tree. The simple set of logic rules
and symbols within the tree structure make a qualitative and quantitative evaluation
of complicated systems. Fault trees are simple to design, but it is difficult to solve
complex tree structure. All the features of fault tree and additional capabilities are
exploited by the Attack trees.
Like fault trees, Attack trees (AT) are also representations of reality, providing a
simplified representation of complex real-world drivers. The accuracy underlying
the drivers and future analysis is determined by the time or effort spent in learning
and norms made.
The attacks on the target is represented by an upside-down tree structure with the
goal as the root node, the sub-goals are different ways of achieving that goal and
leaf nodes as the lowest level tasks. The leaf nodes contain user-defined values
called indicator values to store attributes of that leaf node.
Leaf nodes can have Boolean value (true/false), explicit value (1-low to 4-high),
or continuous value (cost: 0 to any dollar amount). There can be additional options
for continuous node too. In the attack tree, some part of the openly accessible attack
data is used as indicator values. A complex tree can have numerous attack scenarios
only if all possible paths are covered to reach the root. A threat agent profile helps
to reduce the number of tree attacks. Attack tree consists of physical/cyberattacks
and illustrates touch points amongst them.
366 S. N. Shinde and R. P. Shinde
This work deals with communication security aspects of DNP3/SCADA. Our work
enhances the security of the DNP3 protocol to alleviate the threats. The most
important focus is on the redistributes the bytes of the protocol, on the augmen-
tation of the CRC algorithm and Blowfish algorithm for better security aspect.
Traditionally DNP3 protocol has only CRC is used for detecting transmission error
[10, 11]. We provide the security in following ways:
• Encrypting DNP3 Packet.
• Modify the internal structure of protocol.
Out of 292, it uses 34 bytes of the DNP3 link protocol data unit for integrity and
security. We redistribute these bytes to enhance the payload range and security of
the DNP3 protocol with the help of following rearranges fields:
• New LH Header
• Sequence Number
• Original LH Header
• Payload Data
• Enhanced CRC
In this proposal, the message is protected by encryption using Blowfish and our
algorithm for authentication as shown in Fig. 1.
• In this ZA protocol, the message is protected by the following two algorithms.
• Blowfish encryption algorithm: Blowfish provides the confidentiality to the data
by encrypting the data.
• Enhanced CRC algorithm for data authentication in the protocol: CRC helps in
the authentication of the data at both sides. In the enhanced CRC technique, it
uses only 4 bytes of DNP3 protocol.
To calculate the CRC, affix M zeroes next to LSB and then divide the affixed
message by G(z). Then correspondingly:
From the Eq. (2) and congruence property, CRC computation on segmented
message (1):
Here, the modulo of Wi(z) by G(z) will be Wi(z), i.e., Wi ðzÞjGðzÞj ¼ Wi ðzÞ.
The operation and are Galois Field multiplication and addition over GF
(2M), respectively.
Now enhanced CRC algorithm is presented as:
(i) Put original message of N bits and divide it into n chunks
[Wn−1Wn−2 … W1W0] and for every chunk size is M bit (N = nM).
(ii) Initially take generator polynomial G(z) and its degree M and at the same
time calculate a coefficient (as discussed above).
(iii) Perform the n-pair Galois field multiplication in parallel and then XOR the
products which give the CRC result.
Here original message divided into small 68 chunks of 4 bytes. These chunks are
undergoing CRC algorithms to provide message authentication. It is used to provide
integrity of the message in SCADA protocol. Here 4 bytes are used in the CRC out
of 20 bytes and the remaining bytes are reserved for future work. So the protocol is
providing authentication and integrity by using blowfish and enhanced CRC.
Through this approach, we rearranged bytes of the DNP3 protocol to enables
confidentiality, integrity, and authenticity. In this protocol, we have done modifi-
cation in the protocol to reserve the bytes and provide security. The Payload data
and original LH header are encrypted by Blowfish algorithm to provide the con-
fidentiality to the message. The 4 bytes of our proposed CRC is used to provide
message authentication.
6 Conclusion
References
1. Saquib Z, Patel D, Rajrajan R (2011) A configurable and efficient keymanagement scheme for
SCADA. Int J Res Rev, June 2011, 1(2):16–24
2. Saxena A, Pal O0, Saquib Z, Patel D (2010) Customized PKI for SCADA systems network.
Int J Adv Networking Appl 01(05):282–289
3. Mahboob A, Zubairi J (2010) Intrusion avoidance for SCADA security in industrial plants.
Collab Technol Syst (CTS), Proc. CTS 2010, 447–452, IEEE Digital Library
4. Bhagaria S, Prabhakar SB, Saquib Z (2011) Flexi-DNP3: flexible distributed network
protocol version 3(DNP3) for SCADA security. Recent Trends Inf Syst, 293–296, 21–23 Dec
2011
5. Saiwan S, Jain P, Saquib Z, Patel D (2011) Cryptography key management for SCADA
system an architectural framework. Adv Comput Control Telecommun
6. Dawson R (1997) Secure communication for critical infrastructure control system. University
of Queensland
7. Majdalawieh M, Parisi-Presicce F, Wijesekera D (2006) DNPSec: Distributed network
protocol version 3 (DNP3) security framework. Adv Comput Inf Syst Sci Eng, 227–234,
Springer, Dodrecht
8. DNP3 Application Note AN2003-001, http://www.dnp.org/
9. Distributed Network Protocol (DNP3). In: IEEE Standard for Electric Power Systems
Communications 2012
10. Rogaway P, Bellare M, Black J (2006) OCB a block—cipher mode of operation for efficient
authenticated. ACM Trans Inf Syst Secur
11. Bellare M, Rogaway P (1994) Entity authentication and key distribution. In: Advances in
cryptology (CRYPTO’93). Lecturer notes in computer Science. Springer, Berlin
12. Schneier B (1994) Description of a new variable-length key, 64-bit block cipher (Blowfish),
fast software encryption. In: Cambridge security workshop proceedings. Springer, Berlin,
pp 191–204, Vol 809, FSE 1993, Lecture Notes in Computer Science
13. Ji HM, Killian E (2002) Fast parallel CRC algorithm and implementation on a configurable
processor. In: IEEE 2002, vol 3
14. Feldmeier DC (1995) Fast software implementation of error detection code. IEEE/ACM Trans
Networking, IEEE/ACM Trans. on Networking, 3(6), Dec 1995, 640–651
15. Joshi SM, Dubey PK, Kalpan MA (2000) A new parallel algorithm for CRC generation,
communication. In: ICC IEEE international conference
16. Sarwate DV (1988) Computation of cyclic redundancy checks via table lookup. Commun
ACM 31(8)
k-Barrier Coverage-Based Intrusion
Detection for Wireless Sensor Networks
1 Introduction
Section 5 discusses the various results and their analysis. Finally, Sect. 6 concludes
the paper.
2 Related Work
probability using k sensing and single sensing in a given WSNs. Moreover, it has
been shown that the amplitude and frequency of a sinusoidal course affect the
probability of intrusion detection significantly while the change in phase value has a
negligible effect on multiple-sensing and single-sensing detection. Hence, the
author concluded that in order to avoid detection by sensors, intruder should follow
a sine-curve path.
In this part, we explain the mobility and network model and discuss the concept of
barrier coverage to evaluate the performance of networks having mobile sensors.
We assume a two-dimensional rectangular area in which an MSN is deployed
having S(A) mobile sensors as shown in Fig. 1. This region is considered to have an
area A and width W, and initially, it is assumed that the mobile sensors are spread
through uniform distribution independently. Under random deployment consider-
ation, the location of sensors can be calculated with the help of a stationary Poisson
process in two-dimensional having density gA . Therefore, the total number of
sensor in the region A will be given by
Various mobility models are available which defines the motion of sensors; here it
is assumed that the motion of sensors is independent on each other and having no
coordination between them. Speed and direction of motion of sensors characterize
their movement. Sensor chooses its direction of motion randomly from ; 2 ½0; 2p
according
tothe distribution with PDF Pr ð;Þ. Sensor also chooses its speed from
ts 2 0; tmax
s according to PDF Pts ðts Þ. A sensor travels to walls of the region with
selected speed and direction after reaching the edge, the sensor bounces back and
chooses another direction, this process keeps repeating. This model is referred to
the mobility model with the random direction [18]. It is assumed that an intruder
follows a crossing path from one parallel boundary to another which is taken as a
line segment.
It is assumed that rs is the sensing range of each sensor. Each sensor senses the
surrounding area and detects the events within its range of sensing. In this work, an
intruder is considered a point that must be discovered by sensors as it penetrates the
boundary. An intruder is considered to be detected by a sensor, when it has been
found within the sensing range of sensors. This model is the basic sensing model
for a sensor and is known as disc-based sensing model.
In this section, the coverage of the intruder traveling path is defined in terms of
k-barrier coverage. A sensor network is considered to render k-barrier coverage,
when the intruder track traversing the breadth of the area is cumulatively discovered
leastwise k moving sensors. The mobile sensor network performance is measured in
terms of k-barrier coverage probability, i.e., Pr ðD kÞ. Where Δ is the cumulative
coverage count by moving sensors for any intruder way. Another coverage
parameter known as the uncovered distance is defined as the mean distance traveled
by the intruder between the consecutive coverage. In addition to the uncovered
distance, the frequency of coverage is defined as sensor coverage count per unit
time (coverage rate) and is also computed dividing the intruder speed by uncovered
distance.
deduced from the kinetic theory of gas molecules. The mean distance traveled by an
intruder among consecutive coverage is denoted by mean uncovered distance Cs .
The k-barrier coverage probability Pr ðD kÞ in mobile sensor networks is achieved
by modeling the sensors coverage rate hs and the uncovered distance kd .
Let us assume that sensor is stationary having sensing range rs initially, coverage
cross section can be formulated with the help of a circle of radius rs as shown in
Fig. 2. The mean uncovered distance can be computed by dividing intruder-covered
distance by the number of sensor coverage, or it can be computed by dividing the
intruder speed by coverage rate Cs .
prs2
C ¼ 2rs þ ð2Þ
ti t
1
kd ¼ ð3Þ
prs2
gA 2rs þ ti t
pr2
where C ¼ 2rs þ ti ts is the stationary sensor’s coverage cross section and gA is
sensor density respectively. Average uncovered distance in a MSN will be calcu-
lated by modeling the mean relative velocity of moving sensors with respect to
intruders. The relative velocity is formulated in terms of intruder and sensor
velocity vector as shown in Fig. 3. To calculate the coverage rate of sensor, the
velocity of intruder ðti Þ will be replaced by average relative speed of the mobile
sensors in the area of interest. Mobile sensors trel . Then the coverage of sensor per
unit time is given by gA c trel where C is the coverage cross section between
mobile sensors and intruders.
Theorem 1 Sensors coverage rate is given by
hs ¼ gA C trel ð4Þ
Proof Let us assume that intruder j has some probability to be detected by some
moving sensors i 2 8 for i 2 8Sð AÞ having cross section Ci and density of sensors
gj , hence
Sensors
vi.τ
k-Barrier Coverage-Based Intrusion … 379
Intruder
X
hs ¼ trel i28SðAÞ
gA Ci
hs ¼ gA :C:trel
Uncovered time duration is the inverse of coverage rate.
Since the k-barrier coverage probability depends on coverage rate of sensors and
uncovered distance.
Theorem 2 The k-barrier coverage probability is given by
Proof Suppose each sensor has coverage rate ðhs Þ, then the probability that an
intruder will have n sensors coverage exactly on its traveling way with breadth W is
given by
An intruder speed relative to moving sensors changes with the angle between their
respective motion directions as shown in Fig. 3. The entire moving sensors move
randomly in all possible directions, a fraction d/=2p of them move in the direction
that is with an angle / of the intruder ðti Þ direction. In this section, we calculate the
mean relative speed taking random direction mobility model which is given by
Z2p
1
trel ¼ trel d/ ð7Þ
2p
0
from Fig. 4
vi
k-Barrier Coverage-Based Intrusion … 381
Zp=2 rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
E ð vÞ ¼ 1 v:ðsin /Þ2 d/ ð9Þ
0
4ti ts
v¼ : ð10Þ
ð ti þ ts Þ 2
5 Result Analyses
Fig. 5 Effect of mobile sensor density on k-barrier coverage probability at different values of k
It has been observed that k-barrier coverage probability increases as the ratio of the
velocity of the sensor to the velocity of intruder increases and reaches to its
maximum value at a certain value of the ratio for different k-barrier coverage
requirements as shown in Fig. 6. Theoretical calculations also validate simulation
results as the velocity of the sensors increases they can cover a large area in less
time and provide full coverage to the every intruder crossing path.
6 Conclusions
network parameters such as sensor density, sensing range, the speed of the sensors,
and intruders.
Novelty of this work is that, instead of having variable velocities of sensors, we
considered the constant velocity of every sensor which remains same throughout
the simulation time. In this paper, we simulated various performance metrics in
terms of k-barrier coverage probability. The probability of k-barrier coverage
reaches to unity at a specific value of k as the number of moving sensor increase as
shown in Fig. 5. For example at k = 10, the k-barrier coverage probability reaches
to one when number of moving sensors is 700. Again, the probability of k-barrier
coverage improves as sensor to intruder velocity ratio increases as shown in Fig. 6.
For example, at k = 20 barrier requirement, the k-barrier coverage probability
reaches to unity when the velocity ratio is 2.5. Moreover, for a given number of
sensors, the k-barrier coverage probability decreases as the k-barrier requirement
increase as shown in Fig. 7. For instances, when 200 sensors are spread in the field
of interest, the k-barrier coverage probability reaches to zero at k = 20 sensor
requirement. Hence we can conclude that mobility improves the performance of
sensor networks.
References
14. Liu B, Dousse O, Nain P, Towsley D (2013) Dynamic coverage of mobile sensor networks.
IEEE Trans Parallel Distrib Syst 24(2):301–311
15. Li J, Jiang S, Pan Z (2009) Strong barrier coverage for intrusion detection in wireless sensor
network. In: Proceedings of the second symposium international computer science and
computational technology (ISCSCT’09), pp 62–65
16. Wang Y, Wang X, Xie B, Wang D, Agrawal DP (2008) Intrusion detection in homogeneous
and heterogeneous wireless sensor networks. IEEE Trans Mob Comput 7(6):698–711
17. Wang Y, Leow YK, Yin J (2009) Is straight-line path always the best for intrusion detection
in wireless sensor networks. In: 15th ICPDS, pp 565–571
18. Camp T, Boleng J, Avies V (2002) A survey of mobility models for ad hoc network research.
Wireless Commun Mob Comput 2(5):483–502
Performance Analysis of Vulnerability
Detection Scanners for Web Systems
1 Introduction
2 Related Work
gathered by hidden web crawlers to effectively craft the attack string, which greatly
unfolds the vulnerability present in the system. The search area is increased when
used in conjunction with Access Authentication database table (AADT).
Dessiatnikoff [10] have extended the Xin Wang methodology [9] and suggested
a method to predict the vulnerability present in the code. The authors have classified
the response pages into three clusters based on the distance of normalized change in
their responses. Since a response is always associated with a request and vice versa.
Hence it is enough to check the responses. Based on these clusters, the author is
able to predict the Vulnerability present in the system with a greater accuracy than
other scanners present.
Damjanovic and Djuric [11] have suggested that by using domain knowledge of
a particular web application in conjunction with functional programming paradigm
and Model-Driven Architecture, we can draw the attack model. It is possible to
model the attack using this knowledge in the form of attack tree, with a global goal
of attack the system. Analyzing this tree unfolds the vulnerability present in the
system. Later further gained knowledge through this attack model can also be
incorporated into the system to enhance the method’s effectiveness.
Buja et al. [12] used Four-layer approach to Prevent and Detect a Vulnerabilities
in the web system. In his model, he used Boyer–Moore string matching algorithm
which reduced the time in matching the pattern in brute-force approaches. This
algorithm is mostly used to detect virus Detection where large data is analyzed in a
short time. It improved the detection speed in quite a number.
Many researchers are trying to improve the detection capability of a scanner.
Since many of the best scanners are only having detection capability of around 40%
which is not a very good upper limit. Our goal is to choose scanners such that this
upper limit is increased. This can be done by using problem-specific scanners, i.e.
use of different scanners based on the criteria or problem at disposal.
3 Proposed Work
Since the advent of Web applications, users have now power to do many things
such as E-commerce, Online Banking, etc. For this purpose, application store and
send confidential data to server. Now, since applications are prone to
Vulnerabilities. Hence these vulnerabilities are used to attack on web applications.
Which later results in the compromise of confidential data and manipulation of
system. To obliterate these attacks probability, we tend to find the underlying
vulnerabilities in initial states so that it will not make system Vulnerable in later
stages. With all our best efforts 100% detection of vulnerabilities is not possible. So
one best alternative is that we deploy the web application. After that, we scan it for
potential vulnerabilities and rectify them as we encounter them making it a part of
maintenance. However, many vulnerabilities are only surfaced when they are
deployed in the real world. Also, some vulnerabilities do not even lead to a suc-
cessful attack and hence investing unnecessary power is not a good thought.
390 S. Singh and K. Singh
For this purpose, we use vulnerability detection scanners to detect the vulner-
abilities present in the system. They use the Black-Box approach to detect the
vulnerabilities, i.e., they do not need the code of the application to work.
Web application contains Vulnerabilities which are used to attack the application in
order to compromise the security of system. Vulnerability scanners are used to
detect those vulnerabilities. After successful detection, we can rectify these vul-
nerabilities from the system. Hence making the system more secure. All of
Vulnerability scanners use the following basic procedure to detect the vulnerabil-
ities present in the system.
Figure 1 shows the typical working of a scanner. Every phase uses information
generated by its component to successfully complete the task. Different scanners
deploy these components in a different manner and hence greatly changing the
efficiency of the scanner.
3.2 Tools
In this proposed work, we are trying to set a benchmark for some well known
open-source scanners. Our work is focused on evaluating their performance based
on different parameters such as time taken in scan, Number of Vulnerabilities
detected, etc. For this, we have chosen three well-known open-source Vulnerability
scanners. Table 1 shows the list of Scanners used for this purpose.
These scanners are top in their league with most downloaded scanners by
penetration testing users.
comparison statics for the scanner tools. In this work, we have fed these vulnerable
web applications as Input data to scanner tools. The output results are analyzed to
show the different comparison statistics of vulnerability scanners.
392 S. Singh and K. Singh
using the different scanners available. In this model, we used three open-source
scanners available to users. Applying these scanners to datasets available online we
are able to produce scanner outputs, i.e., vulnerability detection reports for each
data set. After that, the proposed work is applied on these generated reports. By
applying the proposed algorithm to the generated reports, different analyses are
done based on criteria specified. According to specified criteria, a scanner is ranked.
For different criteria, different scanners get the high rank. The final generated
reports show the efficient scanner according to criteria specified.
Proposed Algorithm: In this proposed work, the algorithm works in two dif-
ferent phases. The first phase is to generate scanning reports produced by vulner-
ability detection scanners. Next phase uses these reports to generate the ranking of
scanners. The algorithm in pseudocode is as follows:
will show the efficiency of these scanners time-wise, i.e., which is taking less time
to finish the scan. Algorithm will rank these scanners based on this criteria. Change
in criteria will also lead to change in ranking of the scanners.
This section explains the experiment performed and result obtained. A detailed
explanation is also given for the result obtained.
The system configuration for the experiment to perform is listed below. Some are
Compulsory as to run the scanners used.
The Hardware configuration for this experiment is which allowed successful and
efficient execution of the experiment.
– Processor: Intel Core i7 (3.4 GHz)
– Main Memory:1 GB Minimum
– Hard Disk Space: 20 GB Minimum
– Input Device: Keyboard and Preferred Mouse
Software Requirement for this work is as follows:
– Operating System: Ubuntu 14.04—Tools and IDE used: Latex, Gummi editor.
Using the experimental setup, when the proposed algorithm is run. We obtained
Different results based on the criteria we specified.
First Run: For our first run, we chose the Criteria as “Total number of vul-
nerability detected”. In this run, we are interested to find the scanner which is
capable of generating most number of vulnerability detection alerts.
Figure 3 shows the alerts detected when performing scans on the data sets. An
alert is a potential vulnerability in the web application. It is evident from the figure
that ZAP detected more alerts than the rest of the two scanners. This is because of
two main reasons:
1. Since ZAP includes the Top 10 vulnerability profiles standardized by OWASP,
it is more efficient to detect the vulnerability since it has pre-information about
them.
Performance Analysis of Vulnerability Detection … 395
30
Wa
W3
ZA
25
20
15
10
0
Acu Vicn Web Zero
Data Sets
2. Usually, a request has multiple responses and hence ZAP has multiple alerts for
the same request. Hence, the high number of alert detection.
It may look that ZAP is accumulating unnecessary alerts unlike the rest two. But
in many situations, a vulnerability is often surfaced when certain conditions are met
by generating as many as alerts, ZAP increases its coverage area. So, considering
this point ZAP is able to generate more possible vulnerable points than rest of the
two.
So, for our first criteria, we can see that ZAP has higher performance and hence
higher ranking.
Second Run: For this run, we chose the criteria as “Shortest time taken in one
scan”. When we apply the proposed algorithm, we get the result as shown in Fig. 4.
In this case also ZAP has shown the most efficient performance, which took the
least time in all of the scanners to finish the scan. On an average, ZAP was 167%
faster than w3af and 236% faster than Wapiti. In short, ZAP is efficient in the
manner of time taken to complete a scan. This duration may differ as enabling
different plug-ins affect the time to scan drastically. But the algorithm is able to
show the comparison for each dataset.
It is interesting that Wapiti is taking more time than others. Prime cause of this is
that when Wapiti is executed without any specific setting it runs all its modules one
by one, and hence takes more time whereas W3af and ZAP run only the basic
module configuration. So time duration differs in greatly by the plug-ins and
modules configured.
396 S. Singh and K. Singh
6
Wapi
W3
ZA
5
0
Acua Vicnu Web Zero
Data Sets
Third Run: In this Run, we set criteria as “Severity of Vulnerability”, i.e., High,
Medium and Low severity vulnerabilities. The proposed algorithm generates the
following reports.
Figure 5 shows the report for Acuart dataset [14]. As we can see, in this case
ZAP stands out by generating more number of alerts in each section. Whereas
Wapiti and W3af have comparatively the same performance.
Figure 6 shows the same report for dataset 2, i.e., Vicnum project [15]. Which
also shows the same results as in case of dataset Acuart. But what is interesting in
this is that although ZAP generated high number of alerts in other category. The
High-risk factor is almost the same in each of the scanner statistics. If we would
have chosen criteria based on only High-risk Vulnerability, the output would have
ranked all the scanners with the same rank.
90
High Risk
80 Medium Risk
Low Risk
70
60
50
40
30
20
10
0
Wapiti W3af ZAP
Vulnerability Scanners
40
High Risk
35 Medium Risk
Low Risk
30
25
20
15
10
5
0
Wapiti W3af ZAP
Vulnerability Scanners
120
100
80
60
40
20
0
Wapiti W3af ZAP
Vulnerabiltiy Scanners
Figure 7 shows the report for dataset 3, i.e., for Web Scanner Test suite [16].
A similar report has been generated for the dataset Zero Bank [17]. It is shown in
Fig. 8.
These two reports have rather interesting factors associated with them. For
Fig. 7, if we consider medium risk vulnerability alerts it is zero for W3af, and
High-risk type are very less which may be due to unable to bypass ther authenti-
cation page, which reduced the number of detected vulnerabilities. Since our
algorithm uses the scanner report as input it also affects the working of algorithm,
ranking a scanner incorrectly. Hence for better ranking, it should be desirable to
have a good scanning result. As For Fig. 8 results are changed, as we can see in
other results ZAP is prominently ranked but in this result ZAP is unable to detect a
high-risk vulnerability which in turn is detected by W3af in high number. Hence,
398 S. Singh and K. Singh
25
High Risk
Medium Risk
Low Risk
20
15
10
0
Wapiti W3af ZAP
Vulnerability Scanners
our algorithm would have ranked W3af as efficient if we would have chosen the
criteria only for high-risk vulnerability. These results show that if we choose dif-
ferent criteria different ranking is given to the same report.
5 Conclusion
This work deals the issue of how to prioritize Web vulnerability scanners in dif-
ferent situations. However, these scanners work well in some situation and poor in
some situations. Hence, it is crucial to adaptively use them as the scenario arrives.
The proposed algorithm tries to deal this issue by analyzing the scanner’s output.
By this, we can roughly estimate that, in which criteria which scanner to be used.
We here showed that for the criteria being the “scan time” or “severity of alerts”
will produce different statistics. Based on these, the proposed algorithm tries to rank
vulnerability detection scanners. As with every work, this algorithm also is far from
being perfect. Ambiguous ranking may possible if reports are biased or erroneous.
Algorithm does not have the capability to predict similarity between reports, which
when added, will enable the algorithm to rank scanners more accurately. Future
work includes addition of this functionality to improve the results so that better
results can be obtained.
References
Keywords Multicast communication ECDSA Source authentication
ECCSA Elliptic curve cryptography Hash tree Non-repudiation
1 Introduction
The large-scale development of Internet and use of electronics meant for commu-
nication resulted the new digital era of communication. The data or information can
be sent to various network like unicast, broadcast, multicast, etc. In the case of
Y. Mohan (&)
CSED, NERIST, Nirjuli 791109, Arunachal Pradesh, India
e-mail: [email protected]
C. R. Krishna
NITTTR, Chandigarh 160019, India
e-mail: [email protected]
K. Singh
School of Computer & Systems Sciences, Jawaharlal Nehru University,
New Delhi 110067, India
e-mail: [email protected]
2 Related Works
The literature possess several approaches and models for providing source
authentication in multicast communication. The issues and challenges in the area of
multicast security are described in this section existing multicast source authenti-
cation protocol such as simple off-line chaining, tree chaining, EMSS, and HMSA
are described with their advantages and disadvantages.
In Hash chaining [9] scheme, the working of sender and receiver are described
below into the blocks [15] then the hash of the first block is computed and signs the
hash of the first block. The technique of the hash chaining scheme, sender first
divides message M into 4 blocks {B1, B2, B3, B4} then computes the hash of the first
block, signs it, and transmit to each receiver.
In tree chaining [16, 17] scheme each packet carries the required authentication
information so that each can be individually verifiable. In other words, even if
n − 1 out of n packets are lost the authenticity of the single received packet can be
verified. The stream is signed block by block.
Efficient Multi-chained Stream Signature (EMSS) [18, 19] scheme each packet of
the stream is hash linked [20, 21] to many target packets. Even if some packets are lost;
Performance Evaluation of Multicast Source Authentication … 403
a received packet is verifiable if it remains a hash-link path that relates the packet to a
signature packet. For a given packet, the EMSS chooses target packets randomly.
Jin et al. [22] proposed a hybrid approach (HMSA) in which hash tree and hash
chaining scheme are combined. In this approach, the author has targeted on the
main disadvantage that occurs with both the scheme.
This section explained the existing multicast source authentication protocol with
non-repudiation [11] and their advantages and disadvantages. There is no scheme
which will satisfy all the requirements for multicast source authentication. In the
next section, a novel multicast source authentication with non-repudiation protocol
hash redundancy mitigation scheme for multicast source authentication [23] is
proposed which makes a tradeoff between communications overhead [24] and
robustness [25] against the packet loss.
3 Proposed Method
(a) H118
H114 H158
(b) H218
H214 H258
Fig. 1 a Hash generation process for block 1. b Hash generation process for block 2
Receiver Side
To verify Sender signature (e, d) on H: Receiver associated with public key Ca
does the following:
• Verify that e and d are integers between 1 to (N − 1)
• Generate h = H(m)
• Generate t ¼ ðd 1 Þ mod N
• Generate v1 ¼ hl t mod N and v2 ¼ et mod N
• Generate curve point x v1G þ v2 Ca
• If a = 0, then reject the signature, v ¼ a mod N
• Accept the signature if v = e.
Digest Regeneration and Verification of Root Hash
• Receivers first receive the signed hash root of the first block.
• Receivers unsigned the root hash and store it.
• Receivers receive packet P11 and compute H11.
• Now regenerate the hash root of the first block with help of H12, H134, H158 and
computed first block root hash [H118].
Receivers verify the authentication of P11, H12, H134, H58 and second block hash
of root, if stored root hash of block one is identical with the computed root hash of
block one (Fig. 2).
• Receivers store the value H12, H134, H156; second block root hash [H118].
• Receivers get P12 along with previous packet hash, i.e., H11 then it computes the
hash [33] of packet P11 and generates first block root hash with the help of store
hashes. If computed first block root hash H118 is identical with stored first block
root H118, so source is authentic along with packet P2.
• Same way receiver received packet P3, P4 and with the help of stored value of
hash to generate the first block root hash H118. The proposed work flow chart is
given in Fig. 3.
We use QualNet simulator version 5.0 to simulate our work. QualNet simulator
provides wide a variety of simulations
A platform that can predict wireless wired and mixed platform network and
networking device performance.
406 Y. Mohan et al.
There are following parameters used for implementation of the work are shown in
Table 1.
The general scenario of multicast is shown in Fig. 4. In this topology, there is one
source and there are eight receivers. Source needs to send packet only once then in
the network cloud there are many numbers of routers which makes a copy of the
Performance Evaluation of Multicast Source Authentication … 407
packet and send to its neighbor routers. Finally, packet is reached to the end router
which makes many copies of packet as the number of receivers in a particular
multicast group then transmit the packet to that entire receiver.
There are many schemes [34, 35] discussed in literature survey and they used the
RSA for source authentication for multicasting. According to NIST recommenda-
tion, achieving 128-bit security means that the RSA key should be at least 3072 bits
although the same security can be provided using Elliptic Curve Cryptography
Performance Evaluation of Multicast Source Authentication … 409
Digital Signature Algorithm (ECDSA) [36] with the key of 256 bits. Hence the key
size has been reduced.
Effect of Packet Size on Computation Time
It can be observed from the Fig. 5 that the computation time in case of HTS is
highest and computation time SHS and HTSS are approximately equal and greater
than ECCSA scheme.
Effect of Packet Size on Computation Time
It can be observed from the Fig. 6 that the computation time in case of HTS is
highest and computation time SHS and HTSS are approximately equal and greater
than ECCSA scheme.
Effect of Packet Size on Communication Overhead
It can be observed from the Fig. 7 that the communication overhead is less than
SHS, HTS, and HTSS.
Effect of Packet Size on Verification Rate
It can be observed from the Fig. 8 that the verification rate of ECCSA is greater
than the HTS [37] but less than HTSS and SHS. The verification rate is a little bit
less but the other advantage of ECCSA schemes is less communication overhead
because the ECCSA scheme did not send the redundant data through the channel.
Effect of Packet Size on Communication Overhead
5 Conclusions
References
Author Biographies
Yogendra Mohan has completed his ME (Computer science and Engineering) and is currently
working as Assistant Professor in the Department of Computer science and Engineering, North
Eastern Regional Institute of Science and Technology (deemed to be university—MHRD,
Government of India), Nirjuli, Itanagar, Arunachal Pradesh, India. Before he joined NERIST, he
worked as Assistant Professor in various colleges of AKTU, Lucknow, for more than 10 years. He
also worked as software developer for 2 years. His areas of research are computer network security
and cloud computing.
Dr. Rama Krishna received B.Tech. from JNTU, Hyderabad; M.Tech. from Cochin University of
Science and Technology, Cochin; and Ph.D. from IIT Kharagpur. He is Senior Member, IEEE,
USA. Since 1996, he is working with the Department of Computer Science & Engineering,
National Institute of Technical Teachers’ Training & Research, Chandigarh, and currently holding
the position of Professor and Head. His areas of research interest include computer networks,
wireless networks, cryptography and cyber security, and cloud computing. To his credit, he has
more than 80 research publications in referred international and national journals and conferences.
He acted as Associate Editor for International Journal of Technology, Knowledge and Society. He
is a member in advisory/technical committees of many national and international journals and
conferences and also chaired many technical sessions. He is a reviewer of Elsevier Journal of
Vehicular Communications, Elsevier Journal of Computers & Security, Elsevier Journal of
Information and Software Technology. He has 20 years of experience in organizing more than 100
training programs in the upcoming areas of CSE and IT for the faculty of engineering colleges,
polytechnics, and industry professionals. He is instrumental in launching various initiatives at
NITTTR Chandigarh toward paperless office.
Dr. Karan Singh has completed his B.Tech. (Computer Science & Engineering) from Kamala
Nehru Institute of Technology, Sultanpur, in 2004 and M.Tech. (Computer Science &
Engineering) from Motilal Nehru National Institute of Technology, Allahabad, UP, in 2006. He
has completed his Ph.D. (Computer Science & Engineering) from Motilal Nehru National Institute
of Technology, Allahabad, UP, in 2010. He has more than 10 years of experience in research and
teaching. Currently, he is associated with School of Computer and Systems Sciences, JNU, New
Delhi, India. His research areas are computer network and information security. He supervises 23
research candidates. He has more than 40 research papers (journal, IEEE conferences, national and
international conferences) and 2 are accepted. He is the reviewer of conference and journal papers.
He worked as General Chair of 9th International Conference, QShine 2013. He was an organizer of
the workshop with ICUM Conference, Russia, and trying to open a research platform in India. He
had taught more than 10 subjects to PG/UG classes. He was involved in many administrative
activities. He has designed a computer laboratory. He is a professional/life member of various
bodies such as Association for Computing Machinery (ACM), New York; Computer Science
Teachers Association (CSTA), USA; Computer Society of India (CSI), Secunderabad, India;
Cryptology Research Society of India (CRSI), Kolkata, India; Institute of Electrical and Electronics
Engineers (IEEE), USA; International Association of Computer Science and Information
Technology (IACSIT), Singapore; Institute for Computer Sciences, Social-Informatics and
Telecommunications Engineering (ICST), America; International Association of Engineers
(IAENG), Hong Kong; Association of Computer Electronics and Electrical Engineers (ACEEE),
India; Internet Society (ISOC), USA; Academy & Industry Research Collaboration Center (AIRCC).
Design and Implementation of a Secure
Hierarchical Trust Model for PKI
1 Introduction
Organizations need enhanced security for data and strong credentials for identity
management. Digital certificates are used for secure data and proper authentication
from users and computers both within and outside the organization. Most com-
monly used certificates are Digital Certificates, which are part of the public key
infrastructure (PKI). A PKI is the combination of software, hardware, key gener-
S. Tanwar (&)
Department of CSE FET, Mody University of Science & Technology,
Laxmangarh, India
e-mail: [email protected]
K. V. Prema
Department of CSE, Manipal Institute of Technology, MAHE, Manipal,
Karnataka, India
e-mail: [email protected]
2 Trust Model
A trust model provides a framework for building a trust relationship among the
entities. The implementation of a PKI requires ensuring the trust relationship among
the end entities for a secure communication over the unsecured channel. The
awareness of the trust relationships leads to the establishment of a trust model that
the PKI enforces [2].
The hierarchical trust model is like an upside-down tree structure, root is the
starting point of trust [3]. All nodes of the model have to trust the root CA, and keep
a root CA’s public-key certificate [4]. The Root CA’s self-signed certificate is used
for signing other CA certificates and its subordinate CA’s certificate. It can be a
public trusted company such as Verisign.
In hierarchical trust model, CAs are assembled under a common root CA, which
issues certificates to Sub CAs. The hierarchy can have an arbitrary number of
levels, usually, it has two levels: Root CA and certificate issuing CAs [5]. It has a
single root CA and is holding all certificates; all end-users refer to and trust it for all
transaction. Hierarchical Model can also have Registration Authorities (RAs) which
are the initial processing points of user’s identification and issues key pairs. RA
produces flexibility for smaller groups by allowing them to have their own local and
customized services.
(a) Root CA
Root CA is trust anchor for all the users like Controller of Certifying
Authorities (CCA). All nodes have to trust the root CA, and keep the root CA’s
public key certificate. Root CA do the cross certification between two users to
communicate. It generates certificates for the intermediate/Sub CAs, which in
turn generates certificates for the leaf CAs, and the leaf CAs generate certifi-
cates for the end entities (users, network devices, applications). Root CA is
self-certified and generates certificate for the CA containing the entity’s identity
and public key [6]. The generated certificate is signed by the Root CA. Its
public keys must be distributed to all entities that trust on its certificate. The
level of trust that a Root CA has depends on the level of acceptance that other
entities have in that Root CA [2].
Design and Implementation of a Secure Hierarchical Trust Model for PKI 417
3 Problem Statement
A certificate represents a trust from the CA to the owner of the certificate [8]. The
advantage of the hierarchical structure is a short and definite path and is easily
traceable back to a trusted node. Our main objective is to integrate security prin-
ciples in a hierarchical model. This model implements authentication (digital cer-
tificates), confidentiality (encryption), and integrity (SHA 512) non-repudiation
(digital signature).
Root CA (level 0)
PKI CERTIFICATE CHAIN
4 Proposed Approach
5 Implementation
This approach is implemented on java jdk1.7 version with Xampp for My SQL and
Apache Server. Digital Certificate is a most important module, which is signed by
CA/Sub CA to sign the certificates and request. For generating digital signature we
used a hybrid approach of cryptography.
Symmetric key encryption—Message Digest (MD) is calculated using an
instance of SHA-512. Then MD is encrypted with a shared key. For security
purpose, the shared key is encrypted with the receiver’s public key so that key
cannot be decrypted by anyone else who does not possess the matching private key
and also ensure that the public key is associated with the user.
CA-1 CA-2
User User
Internet
Root-CA
CA1 CA2
1. Root CA is self-certified by generating aself-signed certificate by using RSA for key generation.
keyPairGenerator.initialize(2048);
privkey = pair.getPrivate();
pubkey = pair.getPublic();
2. Sub CA send certificate request to Root CA, so that it will be certified by Trusted Root Server like
Controller of Certifying Authorities (CCA).
3. String
concat1=srno+oo.getFname()+oo.getLname()+oo.getCity()+oo.getState()+oo.getOrgunit()+oo.getOrg()+oo
.getEmail()+oo.getCountrycode()+oo.getMobileno()+oo.getIssuedBy()+oo.getIssuedTo()+oo.getValidto()+
oo.getValidfrom()+publickey.toString();
System.out.println("Fingerprint="+msgdigest);
msgdigest=new sha512().sha512(concat1);
oo.setDigest(msgdigest);
byte[] b=new byte[1024];
b=concat1.getBytes();
msgdigest=sh.new Digest().digestIt(b);
4. Root CA Generate symmetric key by creating the instance of AES-128 bit algorithm.
5. Encrypt the hashed data which is send by the Sub CA in the form of certificate request with symmetric key
that was generated in the step 4.
E [ ( + − )]
6. Encrypt the symmetric key with public key of CA, so that only recipient that is having private key
(private/public key pair) can only decrypt the symmetric key.
E [ _ ]
6 Results
Each CA and sub CA have a public key database in which all the public key of the
issuer and issue is stored in the form of an object by taking its data type as BLOB
(Figs. 4, 5, 6, 7 and 8).
Design and Implementation of a Secure Hierarchical Trust Model for PKI 421
7 Conclusion
The hierarchical trust model designed by us is secured than the existing one. We
have created a digital certificate for 2048 bits and use the hybrid approach for
generated the digital signature. As the use of encryption and then storage of Java
objects in BLOB enables confidentiality and message integrity. Hence, the system
will enable secure communication and provide proper authentication. Password,
private key, issuer public key, message and other sensitive information is stored in
the BLOB form and hence is highly secure both from client’s side attack, back-end
attack and also during transmission over internet. The application is completely
based on OOPs concept and hence can be implemented in any kind of organization
may it be academics, government, public sectors, banks, etc.
8 Future Work
This model is implemented in Java. One can implement it with more public key size
such as 4096 bits and SHA 1024 bit key. It would be more secure as the long key is
secure. One can also simulate it on simulators such as MATLAB or OPNET
simulator. The model can then be applied to two distinct working organizations
having separate policies. Policy mapping rules can be made to verify certificates of
each other.
References
8. Audun J (2013) PKI trust models. In: Theory and practice of cryptography solutions for secure
information systems
Author Biographies
1 Introduction
Every person stores huge amounts of data like e-mails, contacts, calendars, docu-
ments, photos, and on the net. To cover and protect the privacy of online delicate
data is another system. This requires that you know which computers will be
attached to each other so that the key can be present on each one. It is same as a
secret code that each of the computers must know in order to translate the
information [1].
Java: It is one of the most robust, mainly used, and perfect programming lan-
guages for creating enterprise applications. Over the years, Java development has
evolved from applets run on a web browser (Chrome, Mozilla) to large enterprise
distributed applications run on multiple servers. Presently, Java has three different
flavors, and each addresses certain programming requirements [2].
Encryption: A practice of changing simple text into secret message text is called
as Encryption. Encryption technique is used by cryptography to send secret mes-
sages through at mid-channel. The encryption processes require two parts key and
algorithm.
Decryption: It is just an antipole process of encryption of text.
Plain Text: It is the main message that somebody wants to broadcast with the
other end is mentioned as plain text. For example, Neeraj sends “Dear Ankur,
Welcome” message to the Ankur. In this, “Dear Ankur, Welcome” message
showed.
Cipher Text: The message that is not easily known or useless is what treated as
cipher type text. By the technique of cryptography, the early message is converted
into non-readable form before the broadcasting of the actual message. Like
“Ank172#@81ukl8*^5%” is a cipher text show “Hello Ankur how are you” [3].
Hashing: It is used to build, search or delete data from the hashtable. Hashing is
so commonly used in computing that one might expect that there is no dearth of
programmer to understand well hash functions and that choosing a suitable function
not be a difficult task [4] (Fig. 1).
The symmetric key uses the similar key for message encryption and decryption
process; still, a message can include a distinct key than others. The drawback of
symmetric cipher is the central management necessary to use them soundly. Any
specific pair of conveying parties need, elegantly, shares a distinct key, and may be
each of the cipher text. The number of keys needed increases as the second power
of the number of network members, which very quickly need critical key man-
agement scheme to controlling form and secret keys. The Diffie and Hellman’s
research sparked widespread efforts to finding public-key encryption [7] (Fig. 3).
430 A. Saxena et al.
2 Review of Literature
This encryption technique depends on the capacity across the shortcut encounter,
which is more active than the brute force process. Along the block duration and 128
bit key length for AES algorithm, it has not found to six or even more trolls on an
interpreted version of the usage of shortcut encounters [8].
User study three sets of initial keys only one bit difference after ten round
expansions with the round key technique which chosen by random, in Table 2
through them, the first set key has variation in 16th built, the second in 128th, and
last in 40th. The big divergence of this key variation after ten rounds is show in
Table 3 [9].
3 Methodology
The encryption user must use a secret key with a technique. In this process, user
uses a technique called advanced encryption standard 128 and the string bytes
“AnkurSaxena” as the secret key. Advanced encryption standard technique can use
a key of 128 bits, so programmer selected that key.
The following output is got from the above test; user clearly shows that the
original text is replaced with decryption:
Plain Text: Saxena
Encrypted: sbhCap4urE50a/d
Decrypted: Saxena (Fig. 5)
4 Conclusion
Encryption technique plays a key role in network security. This work reviewed the
competence of encryption, AES techniques, AES with Java over tomcat web server
or application host. Roots on the text files are used and the technical outcome, it
was achieve that AES technique utilize least encryption. It is also observed that
decryption of AES technique is safer than various other techniques. By using
simulation result, it is calculated that AES algorithm is a better than DES technique.
Our future research will target on compared and analyzed cryptographic encryption
like AES, DES. It will incorporate experiments on audio, image, and video
encryption or decryption, and the target will be to advance encryption and
decryption velocity or time.
References
1. Saxena A, Jakhmola R (2011) Securing confidential data using Java/J2EE. Int J Sci Technol
Manag 2(3):54–59
2. DSarkar D, Jaiswal A, Saxena A (2015) Understanding architecture and framework of J2EE
using web application. Int J Comput Sci Inf Technol 6(2):1253–1257
3. Thambiraja E, Ramesh G, Umarani R (2012) A survey on various most common encryption
techniques. Int J Adv Res Comput Sci Softw Eng 2(7):226–233
4. Saxena A, Chaurasia (2014) Key and value paired data using java hash table. Int J Eng Manag
Res 4(1):81–89
5. Mahajan P Dr, Sachdeva A (2013) A study of encryption algorithms AES, DES and RSA for
security. Glob J Comput Sci Technol Netw Web Secur 13(15):15–22 Version 1.0
6. Prashant G, Deepthi S, SandhyaRani K (2013) A novel approach for data encryption standard
algorithm. Int J Eng Adv Technol (IJEAT) 2(5):264
7. Agarwal V, Agarwal S, Deshmukh R (2014) Analysis and review of encryption and
decryption for secure communication. Int J Sci Eng Res (IJSER) 2(2):1–3
8. FIPS PUb 197-the official AES standard. http://www.techheap.com/cryptography/encryption/
fips-197.pdf
9. Chen† Q, Tang Z, Li Y, Niu Y, Mo J (2011) Research on encryption algorithm of data
security for wireless sensor network. J Comput Inf Syst 7(2):369–376
10. Bhati S, Bhati A, Sharma SK (2012) A new approach towards encryption schemes: byte—
rotation encryption algorithm. In: Proceedings of the world congress on engineering and
computer science 2012, vol II WCECS 2012, October 24–26
436 A. Saxena et al.
Author Biographies
1 Introduction
network, the interaction takes place among the wireless nodes and few access
points. Ad hoc network is a type of infrastructure less and decentralized type
wireless network which basically means, there is no actual infrastructure such as
router devices or access points in wireless networks. In routing process, each node
involves itself by forwarding data to and for all the nodes [1].
In ad hoc network, the regulation of which node to forward data is made
dynamically on the basis of network design and connection. Essentially, it is a
network which is generally used in emergency situations. A fixed infrastructure is
not required such types of networks. Nodes which are in close radio range, interact
directly which each other using the wireless links whereas the nodes which are far
from each other take the help of intermediate nodes so that relay message can be
passed. Wireless networks are the networks which make use of radio waves or
microwaves in order to establish interaction between the devices. In such network,
all the nodes act as router.
MANET is mobile ad hoc network. It is self-establishing network which is
infrastructure less in nature. In MANET, different mobiles are associated through
different wireless links. Every mobile node can freely move, which further means
that there is no central control available. In MANET, mobile nodes can join or leave
the network at any instance [2]. MANET is used in some crucial applications such
as emergency salvage, vehicular network, military, and law prosecution. There are
various problems in MANET like security concerns, transfer issue, etc. Due to same
reason, there are different types of attack which are provoked in MANET. These
attacks can be of different types, such as:
• Eavesdropping is a type of attack which takes place in the mobile ad hoc
networks. Eavesdropping is executed to obtain any information which is secret
in nature and is kept classified during entire communication.
• Gray-hole attack’s other name is routing misbehavior attack. It leads to message
dropping.
• Replay attack is a type of attack in which the attacker executes a replay attack
that is repeatedly retransmitted. The actual data that has been captured by the
network is repeatedly transferred. This attack spots the route novelty and brings
out the poor security design.
To isolate these attacks from interaction path in the network, there are different
techniques which we will be discussed further. In this proposed mechanism, we
tried to prevent these attacks (replay attack) by using mutual authentication among
the nodes in the entire network. For this, we used authentication-based protocol
called ALARM using cryptographic mechanism of digital signature in the wireless
sensor network.
Detection and Removal of Security Attacks Using … 439
2 MANET
3 Alarm Protocol
recommended because this serves the purpose of both authentication and security. It
is also to prevent the network from the active and passive attacks.
This basically follows two schemes, i.e., initialization and operation.
A. Initialization
• The group manager is the head of the entire network. He is the one who
adds all the nodes in the network as the group members. During this phase,
every group member is assigned a private key that is unknown to anyone.
This key is required to implement the valid group signatures for security
purpose [3]. Every group member has a public key as well which is only
known to the group manager. The group manager is only responsible for
every group signature and verifies all the signers.
• The group manager is responsible for adding or deleting the group member
[4]. The GM must check whether joining or joining is feasible for the
network or not.
B. Operation
• The time duration is divided into equal parts. While beginning process,
every node member generated a temporary public–private key combination.
• Every stop will let us know about the location of the node through GPRS.
• The GPRS would contain its location, time stamp as well as the temporary
public key.
• When a new “Location Announcement Message” is received, every node
member will check that the same LAM has not been received by them
before [5]. When this is verified, the time stamp with group signature is
checked. If all the entities are verified, the node forwards the LAM to its
neighboring node.
• Whenever a node wishes to interact with the other location, it asks if the
other node already exists there or not and generates a session key if there is
no node at that particular location.
• Then, the message is forwarded to the nodes. The path is chosen based upon
shortest path or other path computing algorithms.
4 Proposed Work
The protocol ALARM is used majorly for mutual authentication among the nodes.
Having read the assumptions like location and time, we get to know that clocks of
the mobile nodes are weakly synchronized. When the clocks are weakly synchro-
nized in any network, then the possibility of replay attack becomes more, making
the data transmission among the nodes very unreliable. In this work, we will isolate
the replay attack in the mutual authentication using ALARM protocol in wireless
sensor network. Using NTP, we can ensure strong clock synchronization among the
Detection and Removal of Security Attacks Using … 441
nodes. The term “strongly synchronized” refers to that if the data is transferred from
one node to other, the processing speed is very fast [6]. If there exists trust rela-
tionship among the nodes, no replay attack is possible in the network because there
is no waiting time for the data transfer during communication.
Due to weak synchronization, the confidential information from the network
may be lost. But while using NTP, mutual authentication among the nodes takes
place and malicious node is removed from the network (Fig. 1).
Fig. 1 Flowchart of
proposed methodology
442 S. Rawat et al.
5 Result
In this figure shown below, it can be seen that the flood messages move to the
monitor node which then identifies the malicious node and finds the best suitable
path for further data transmission (Fig. 2).
In the figure shown below, it is seen that the source node gets reply message
from each node for carrying forward the data communication. In this way, the
interaction among the nodes is stopped, and a new path is established (Figs. 3, 4).
In the figure shown above it is clear that, due to the new proposed algorithm,
whenever a malicious node is detected in the network, we find the best suitable path
hence removing the malicious node from the entire network.
Packet loss:
The two axis of the graph represents different entities; X being the time and Y being
the packet loss. As in this network, the replay attack takes place resulting in higher
packet loss due to delayed transmission of the data packets. The graph basically
represents that earlier there was huge packet loss which is in green color and now it
has been majorly reduced by which is depicted by red line. This is only possible due
to the isolation of the malicious node (Fig. 5).
Network throughput:
Throughput can be defined as the average rate per unit time. In this case, it can be
defined as the average rate of packets delivered successfully in per unit time. The
throughput of any network should always be high. But in this case, due to replay
attacks, it is very low. On the contrary, we see that the network throughput
increases rapidly after isolation of the malicious node. In this graph shown below,
the red line represents the throughput of the network while replay attack is being
taking place. Whereas the green line depicts the new throughput which is very high
after the isolation of the malicious node (Fig. 6).
In this work, we can conclude that due to major properties of the mobile ad hoc
network, various attacks are possible. These properties are open channel, infras-
tructure less network, and variably changing topologies. These attacks can be
prevented by different authentication protocols. In our work, different types of
attacks and their detection, isolation, and impacts on the network are well analyzed.
The main aim of security is that the packet transmission in the entire network from
one location to another should be reliable and verified. In the network, all the nodes
should follow strong trust relationship because any type of malicious node can
444 S. Rawat et al.
References
1. Levya Mayorga I (2014) Performance analysis of a non preemptive hybrid WSN protocol in
mobile environment. In: 28th international conference on advanced information networking
and applications workshop, IEEE, pp 486–491, May 2014
2. Shen H, Zhoa L (2013) ALERT: anonymous location based efficient routing protocol in
MANET. IEEE Trans Mob Comput 12(6) June 2013
3. Agarwal P, Ghosh RK (2008) Cooperative black and gray hole attacks in mobile Ad hoc
network. In: 2nd international conference on ubiquitous IMC, Korea
4. El Defrway K, Tsudik G (2011) ALARM: anonymous location aided routing in suspicious
MANET. IEEE Trans Mob Comput 10 Sept 2011
446 S. Rawat et al.
Subhranil Som
1 Introduction
Early computer applications had no or very less security. People understood that
data on computers are tremendously important feature of modern life. That is why
many areas in security began to gain eminence [1]. As outcome researchers are still
at work in the area in cryptography to develop the security more effectively [2].
S. Som (&)
Amity Institute of Information Technology, Amity University,
Uttar Pradesh, India
e-mail: [email protected]; [email protected]
In this paper a new technique has been proposed, where an algorithmic approach
to a compression scheme followed by an encryption of the compressed input
stream, using Elliptic Curve Cryptography (ECC), over prime field, has been cas-
caded. The compression is an Artificial Intelligence (AI) approach where the input
stream is fully read and the repetitive groups in input stream are replaced by some
unused character set. In the encryption scheme, first, the required parameters have
been chosen to satisfying the equation [4a3 + 27b2 mod p 6¼ 0] and a prime number
“p”, which defines the cardinal number of the set. Each “p + 1” elements of the set
are evaluated. Each distinct character in the input stream is reflected to a point over
the elliptic curve, to deduce a point (xA, yA), using an integer value (K) which has
been agreed by both sender and receiver. Receiver of the cipher text by choosing a
point from the set of Generator points generates public key set, using private key
and is distributed among all the senders. This public key set is used in conjunction
with the sender’s private key and the point (xA, yA) to generate the cipher text which
has been passed over to the intended recipient. Receiver takes up the cipher text and
uses private key to find back the point (xA, yA). Lastly, the actual character from
(xA, yA) using K and the original input stream has been evaluated.
In Sect. 2 of this paper, the proposed technique has been discussed, Sect. 3 the
Performance and Analysis are discussed. The conclusion is given in Sect. 4 fol-
lowed by References.
To grow a field through pm elements, it has been indicated as Galois Field [GF(pm)],
polynomial f(x), (degree m) irreducible over GF(p). An elliptic curve (EC) and
finite field GF(q) have chosen in following ways:
(a) Finite field, an efficient illustration of field elements have been chosen in a way
that the processes are easily executed.
(b) The other representations of curve (projective, etc.) have been taken into
consideration.
(c) Number of points on the curve, #E(GF(q)), has been divisible by large prime
“n”.
(d) #E(GF(q)) 6¼ q, where q = #GF(q) (field order).
Encryption Technique Using Elliptic Curve Cryptography … 449
Let us suppose that, “p” is a prime number = 11. Then the cardinal number of the
set of point is (11 + 1).
For x = 0, 1,…, 10, compute z = x3 + x + 6 mod 11
z has been checked whether is a quadratic residue by
Legendre symbol (z/p) = z^(p − 1)/2 mod p = z^5 mod p
If YES, compute two square roots: ±z (p+1)/4 mod p = ± z3 mod p
Hence, the points are: (2,4),(2,7), (3,5),(3,6), (5,2),(5,9), (7,2),(7,9), (8,3),(8,8),
(10,2), O.
Let the generator point in (2, 7) represented by G.
Compute the 2G, 3G, … as follows:
xR = s2 – xP − xQ and yR = s(xP − xR) − yP where s = (yP – yQ)/(xP – xQ)
if P 6¼ Q, s = (3xP2 + a)/(2yP) if P = Q
For user A: User A has chosen random number to generate private key
(Ak) = 7
“Ak” is multiplied with G to obtain the public key as follows:
Ak*G = 7*(1, 5)
This has been done as follows:
G + G = 2G
Hence, 7 * G = 2(2G + G) + G = (7, 2) (Let this point be denoted by GA).
Thus, the total public key pair is [G, GA].
Each user issues their public key.
Like this searching and replacing algorithm carries on till all the repeating group
of characters are replaced. The replacement has been done if the total size of the
repeating string is more than the size of binary number, which replaces the string.
In this section, the proposed algorithm has been analyzed for time complexity
performance in contrast to RSA and Triple-DES. This performance analysis has
been done over text files (*.TXT), and executable files (*.EXE). The analysis has
been done over the parameters “Encryption time”, “Decryption time”, “Character
Frequency”, and “chi-square values”. Java programming languages have been used
for developing the programing implementation of the technique.
Ten different text files sizes have been taken for experiment. The encrypted time,
the decrypted time, and source file sizes have been noted for Triple-DES, RSA, and
proposed technique. Tables 1 and 2 has been shown the encrypted and decrypted
time of growing size of .txt files for proposed, T-DES, and RSA techniques. The
proposed technique has been taken higher time to encrypt or decrypt compared to
Triple-DES and RSA for any file size. Figure 1 shows the pictographic represen-
tation of the same.
Table 1 Comparative study of file size and encrypt time for text files (for proposed, RSA, and
T-DES algo)
File name (.TXT) File size (in bytes) Encrypt Time (in s)
RSA Triple-DES Proposed Algo
adcajavas.txt 629 *0 *0 3.21
license.txt 7168 1 2 21.87
oledbjvs.txt 10,240 1.68 2.56 38.76
nerohistory.txt 17,408 1.99 2.96 40.65
nero.txt 33,792 3.76 4.54 50.43
whatsnew.txt 69,632 6.87 7.32 60.87
new.txt 94,208 7.21 7.98 63.99
c text.txt 132,096 7.97 9.87 65.43
9.txt 540,672 10.65 12.90 68.98
incidia.txt 1,190,912 14.89 15.98 71.65
Encryption Technique Using Elliptic Curve Cryptography … 453
Table 2 Comparative study of file size and decrypt time for text files (for proposed, RSA, and
Triple-DES algo)
File name (*.TXT) File size (in Bytes) Encrypt time (in s)
RSA Triple-DES Proposed Algo
adcajavas txt 629 *0 *0 2.98
license.txt 7168 1 1.9 20.51
oledbjvs.txt 10,240 1.65 2 36.87
nerohistory.txt 17,408 1.99 2.65 39.65
nero.txt 33,792 3.77 3.54 48.87
whatsnew.txt 69,632 6.80 7.01 55.87
new.txt 94,208 7.01 8 60.43
c text.txt 132,096 7.54 9.31 61.08
9.txt 540,572 10.65 13.01 65.72
incidia.txt 1,190,912 14.12 15.99 69.43
Time (Second)
Fig. 1 Encrypt and decrypt time for proposed Algo, RSA, and Triple-DES techniques for .TXT
files
Ten different sizes of .exe files have been taken for experiment. Comparative
studies of time complexity have been done for these files to encrypt and decrypt
using proposed, RSA, and Triple-DES Technique. Experimental results have been
noted down in Tables 3 and 4. From the data of Tables 3 and 4 depicts the pro-
posed technique has taken more time to encrypt and decrypt in comparison to
Triple-DES and RSA technique for any size of the executable files. The picto-
graphic representations of the same have been given in Fig. 2.
454 S. Som
Table 3 Comparative study of file size and encrypt time for EXE files (for proposed, RSA, and
Triple-DES algo)
File name (*EXE) File size (in bytes) Encrypt time (in s)
RSA Triple-DES Proposed Algo
1.exe 28,672 2 4 35
2.exe 96,256 3 6 41
3.exe 130,048 7 9 50
4.exe 175,104 13 17 57
5.exe 292,364 14 19 61
6.exe 355,328 19 23 67
7.exe 613,376 29 34 78
8.exe 775,168 35 41 81
9.exe 1,307,648 49 58 98
10.exe 1,835,003 61 67 102
Table 4 Comparative study of file size and decrypt time for .EXE files (for proposed, RSA, and
Triple-DES algo)
File name (*EXE) File size (in bytes) Decrypt time (in s)
RSA Triple-DES Proposed Algo
1.exe 28,672 2 4 33
2.exe 96,256 4 6 40
3.exe 130,048 8 10 45
4.exe 175,104 15 18 55
5.exe 292,864 16 21 58
6.exe 355,328 21 23 62
7.exe 613,376 33 39 73
8.exe 775,168 41 43 79
9.exe 1,307,645 53 61 91
10.exe 1,835,008 66 72 96
Time (Second)
Fig. 2 Encrypt and decrypt time for proposed technique, RSA, and Triple-DES techniques for .
EXE files
Encryption Technique Using Elliptic Curve Cryptography … 455
Fig. 3 Pictorial representation of chi-square values for RSA, proposed algo, and Triple-DES
4 Conclusion
The main attraction of the proposed algorithm is that, compared to RSA, it provides
more or less similar security for a lesser bits size, thereby dropping processing
overhead. The proposed technique is perfect for constrained situation such as
PDAs, cell phones and smart cards. Though proposed technique has taken more
time for encryption and decryption as compared to T-DES and RSA, it has shown
good result in chi-square test. It has been seen that encrypted file using proposed
algo has high chi-square value. This high value is indicating good security.
References
1. Atul K (2005) Cryptography and network security. Tata McGraw-Hill, New Delhi. ISBN
0-07-049483-5
2. Som S, Mandal JK (2008) A session key based secure-bit encryption technique (SBET). In:
National conference (INDIACom-2008) on computing for nation development, New Delhi,
India, 08–09 Feb 2008
3. Al-Vahed A, Sahhavi H (2011) An overview of modern cryptography. World Appl Program 1
(1):3–8. ISSN: 2222–2510
4. Certicom (2000) Standards for efficient cryptography, SEC 1: elliptic curve cryptography,
Version 1.0, Sep 2000
5. Certicom (2000) Standards for efficient cryptography, SEC 2: recommended elliptic curve
domain parameters, Version 1.0, Sep 2000
6. William S Cryptography and network security, principles and practice
7. Anoop MS (2015) Elliptic curve cryptography—an implementation guide. URL: http://www.
infosecwriters.com/text_resources/pdf/Elliptic_curve_AnnopMS.pdf. Last accessed on April
2015
8. Darrel H, Alfred M, Scott V (2003) Guide to elliptic curve cryptography. Springer, Berlin.
ISBN 0-387-95273-X
9. Henri C, Gerhard F, Roberto A (2005) Handbook of elliptic and hyper-elliptic curve
cryptography. Chapman and Hall/CRC. ISBN: 978-1-58488-518-4
10. Jadhav A (2011) Implementation of elliptic curve cryptography on text and image. Int J
Enterp Comput Bus Syst 1(2): ISSN (Online): 2230–8849. http://www.ijecbs.com, July 2011
11. Kumar R, Jaiswal UC (2011) Experimental investigation of image encryption technique using
public key. Int J Tech 1(1):12–14
12. Sharma RD (2011) Quantum cryptography: a new approach to information security. Int J
Power Syst Op Energy Manag (IJPSOEM) 1(1)
Encryption Technique Using Elliptic Curve Cryptography … 457
Author Biography
Keywords Cross-site scripting (XSS) attacks Script injection vulnerabilities
JavaScript Context-Sensitive sanitization routines HTTP
1 Introduction
S. Gupta (&)
Department of Computer Science and Information System, Birla Institute
of Technology and Science, Pilani, Vidhya Vihar, Pilani 333031, Rajasthan, India
e-mail: [email protected]
B. B. Gupta
Department of Computer Engineering, National Institute of Technology Kurukshetra,
Kurukshetra 136119, Haryana, India
e-mail: [email protected]
that are created on the added features as well as saves such rules in the Rule-Based
Feature Repository module. Once, the initial HTTP response is produced on web
server, it has to pass through the Rule-Based Variance detector, which detects some
deviation among the actual stored features and observed features. Even, slight dis-
similarity detected would be deliberated as suspicious injected code and finally, the
appended features will be extracted from the code of JavaScript before entering into
the second phase of Auto-Context-Sensitive Sanitization. The second phase,
Auto-Context Sensitive Sanitization performs a practice of automated sanitizer
assignment via evaluating and discovering the chunk of injected code. However,
assignment of sanitizer is stationary as well as occasionally deviates to dynamic
whenever it is necessary. The key objective of such module is to determine some
inputs related to sanitization on the path of source code of JavaScript.
3 Discussion
Usually, XSS attacks involve that the web servers, which are controlled by an
attacker, has to initiate a communication with the compromised web pages. Here,
the authors suggested an innovative framework, which is highly inspired by the
current literature works as mentioned in [11, 12]. Our proposed design is reliant
upon the notion of incorporating features of script data in order to discover the
variation between the stored and observed features in the HTTP response. In
addition to this, the notion of incorporating the context-sensitive sanitization
functions has also adopted by our design to detect the XSS attack vectors embedded
in JSP websites. The benefits of our design versus other two recent related XSS
defensive techniques [11, 12] include: (1) it possesses capability for noticing sus-
picious JavaScript function invocation. (2) It is based on the automated prepro-
cessing technique of insertion and abstraction of script features and sanitization
methods embedded in JSP websites. (3) It did not incorporate the valid script as this
could again create the prospect of comprising distant JavaScript file in the related
anchor tag. (4) Our technique consumes tolerable time in detecting XSS attack
vectors as the authors suggested a programmed procedure of insertion of script
features and related sanitization functions.
Our framework simply senses association among the deposited features and
incorporated ones embedded in JSP web platforms. Hence, we could not assure that
our technique can alleviate the attack vectors that circumvent certain conditions that
were present in rule-based repository. In addition, the entirety and correctness of
mined features of script data could not be guaranteed. The authors preferred to
scrutinize the functions for automatic confirmation of retrieved script data features.
However, we considered that our proposed design presents a reasonable assurance
that it will detect certain attack vectors with small percentage of false positives and
incurs tolerable runtime overhead. We will implement the prototype of our design
in Java as a server-side framework. Throughout the course of experimental eval-
uation, we will test and assess the XSS attack recognition proficiency of our
server-side design on the following JSP websites (i.e., JAuction [7]; JVote [8],
MeshCMS [9] and Easy JSP Forum [10]).
server-side design in Java and would assess the suspicious scripts recognition
proficiency of proposed design on JSP websites. In addition, we will also utilize the
HTML5 websites for introducing the concept of inserting the features of script data
and incorporate context-sensitive sanitization routines in their source code.
References
1. Klein A (2002) Cross site scripting explained. White Paper, Sanctum Security Group, June
2. Gupta S, Gupta BB (2016) XSS-secure as a service for the platforms of online social
network-based multimedia web applications in cloud. Multimed Tools Appl 1–33
3. Gupta BB, Gupta S, Gangwar S, Kumar M, Meena PK (2015) Cross-site scripting
(XSS) abuse and defense: exploitation on several testing bed environments and its defense.
J Inf Priv Secur 11(2):118–136
4. Gupta S, Gupta B (2015) PHP-sensor: a prototype method to discover workflow violation and
XSS Vulnerabilities in PHP web applications. In: 12th ACM International Conference on
Computing Frontiers (CF’15), Ischia, Italy
5. Chaudhary P, Gupta S, Gupta BB, Chandra VS, Selvakumar S, Fire M, Goldschmidt R,
Elovici Y, Gupta BB, Gupta S, Gangwar S. Auditing defense against XSS worms in online
social network-based web applications. In: Handbook of research on modern cryptographic
solutions for computer and cyber security, vol 36, pp 216–245, 16 May 2016
6. Gupta S, Gupta BB (2014) BDS: browser dependent XSS sanitizer. Book on cloud-based
databases with biometric applications, In: IGI-global’s advances in information security,
privacy, and ethics (AISPE) series, 31 Oct 2014, pp 174–91
7. JAuction-0.3. http://sourceforge.net/projects/jauction/
8. JVote. Accessed from http://sourceforge.net/projects/jspvote/
9. MeshCMS. http://cromoteca.com/en/meshcms/
10. Easy JSP Forum. http://sourceforge.net/projects/easyjspforum
11. Shaihriar H, Zulkernine M (2011) S2XS2: a server side approach to automatically detect XSS
attacks. In: Ninth international conference on dependable, automatic secure computing, IEEE,
(2011), pp 7–17
12. Shaihriar H, Zulkernine M (2011) Injecting comments to detect javascript code injection
attacks. In: Proceedings of the 6th IEEE workshop on security, trust, and privacy for software
applications, Munich, Germany, July, pp 104–109
13. Gupta S, Gupta BB (2015) Cross-site scripting (XSS) attacks and defense mechanisms:
classification and state-of-the-art. Int J Syst Assur Eng Manage 1–19
14. Gupta S, Gupta BB, Chaudhary P (2017) Hunting for DOM-based XSS vulnerabilities in
mobile cloud-based online social network. In: Future Generation Computer Systems. 12 June
2017
15. Gupta S, Gupta BB (2016) Alleviating the proliferation of JavaScript worms from online
social network in cloud platforms. In: 2016 7th international conference on information and
communication systems (ICICS), IEEE, pp 246–251
16. Gupta S, Gupta BB (2016) An infrastructure-based framework for the alleviation of
JavaScript worms from OSN in mobile cloud platforms. In: International conference on
network and system security 28 Sep 2016, pp 98–109. Springer International Publishing
17. Gupta S, Gupta BB (2016) XSS-immune: a Google chrome extension-based XSS defensive
framework for contemporary platforms of web applications. Secur Commun Netw
9(17):3966–3986
18. Gupta S, Gupta BB (2016) Alleviating the proliferation of JavaScript worms from online
social network in cloud platforms. In: 2016 7th International Conference on Information and
Communication Systems (ICICS), IEEE, pp 246–251
A Robust Server-Side JavaScript Feature Injection-Based … 465
19. Gupta S, Gupta BB (2017) Smart XSS attack surveillance system for OSN in virtualized
intelligence network of nodes of fog computing. Int J Web Serv Res (IJWSR) 14(4):1–32
20. Gupta S, Gupta BB (2016) JS-SAN: defense mechanism for HTML5-based web applications
against JavaScript code injection vulnerabilities. Secur Commun Netw 9(11):1477–1495
Author Biographies
Dr. Shashank Gupta is currently working as an Assistant Professor in Computer Science and
Information Systems Division at Birla Institute of Technology and Science, Pilani, Rajasthan,
India. He has done his Ph.D. under the supervision of Dr. B. B. Gupta in Department of Computer
Engineering specialization in Web Security at National Institute of Technology Kurukshetra,
Haryana, India. Recently, he was working as an Assistant Professor in the Department of
Computer Science and Engineering at Jaypee Institute of Information Technology (JIIT), Noida,
Sec-128. Prior to this, he has also served his duties as an Assistant Professor in the Department of
IT at Model Institute of Engineering and Technology (MIET), Jammu. He has completed M.Tech.
in the Department of Computer Science and Engineering Specialization in Information Security
from Central University of Rajasthan, Ajmer, India. He has also done his graduation in Bachelor of
Engineering (B.E.) in Department of Information Technology from Padmashree Dr. D.Y. Patil
Institute of Engineering and Technology Affiliated to Pune University, India. He has also spent
two months in the Department of Computer Science and IT, University of Jammu for completing a
portion of Post-graduation thesis work. He bagged the 1st Cash Prize in Poster Presentation at
National Level in the category of ICT Applications in Techspardha’2015 and 2016 event
organized by National Institute of Kurukshetra, Haryana. He has numerous online publications in
International Journals and Conferences including IEEE, Elsevier, ACM, Springer, Wiley, Elsevier,
IGI-Global, etc. along with several book chapters. He is also serving as reviewer for numerous
peer-reviewed Journals and conferences of high repute. He is also a professional member of IEEE
and ACM. His research area of interest includes Web Security, Cross- Site Scripting (XSS) attacks,
Online Social Network Security, Cloud Security, Fog Computing and theory of Computation.
Dr. B. B. Gupta received PhD degree from Indian Institute of Technology Roorkee, India in the
area of Information and Cyber Security. In 2009, he was selected for Canadian Commonwealth
Scholarship and awarded by Government of Canada Award ($10,000). He spent more than six
months in University of Saskatchewan (UofS), Canada to complete a portion of his research work.
He has published more than 70 research papers(including 01 book and 08 chapters) in International
Journals and Conferences of high repute including IEEE, Elsevier, ACM, Springer, Wiley
Inderscience, etc. He has visited several countries, i.e. Canada, Japan, China, Malaysia,
Hong-Kong, etc. to present his research work. His biography was selected and publishes in the
30th Edition of Marquis Who’ s Who in the World, 2012. He is also working principal investigator
of various R&D projects. He is also serving as reviewer for Journals of IEEE, Springer, Wiley,
Taylor & Francis, etc. He is serving as guest editor of various Journals. He was also visiting
researcher with Yamaguchi University, Japan in 2015 and with Guangzhou University, China in
2016, respectively. At present, Dr. Gupta is working as Assistant Professor in the Department of
Computer Engineering, National Institute of Technology Kurukshetra, India. His research interest
includes Information security, Cyber Security, Mobile/Smartphone, Cloud Computing, Web
security, Intrusion detection, Computer networks and Phishing.
PHISH-SAFE: URL Features-Based
Phishing Detection System Using
Machine Learning
Abstract Today, phishing is one of the most serious cyber-security threat in which
attackers steal sensitive information such as personal identification number (PIN),
credit card details, login, password, etc., from Internet users. In this paper, we
proposed a machine learning based anti-phishing system (i.e., named as
PHISH-SAFE) based on Uniform Resource Locator (URL) features. To evaluate
the performance of our proposed system, we have taken 14 features from URL to
detect a website as a phishing or non-phishing. The proposed system is trained
using more than 33,000 phishing and legitimate URLs with SVM and Naïve Bayes
classifiers. Our experiment results show more than 90% accuracy in detecting
phishing websites using SVM classifier.
1 Introduction
Phishing is one of the major security threats faced by the cyber-world and could
lead to financial losses for both industries and individuals. In this attack, criminal
makes a fake web page by copying contents of the legitimate page, so that a user
cannot differentiate between phishing and legitimate sites [1]. Life cycle of phishing
attack is shown in Fig. 1. According to anti-phishing working report in the first
Quarter of 2014, second highest number of phishing attacks ever recorded between
January and March 2014 [2] and payment services are the most targeted by these
attacks. The total number of phishing attacks notice in Q1 (first quarter) of
Fig. 1 Phishing life cycle: (1) phisher copies the content from legitimate site and constructs the
phishing site; (2) phisher sent link of phishing URL to Internet user; (3) user opens the link and
fills personal on fake site; (4) phisher steals the personal information of user; (5) phisher deletes the
fake web page
2014 were 125,215, a 10.7 percent increase over Q4 (fourth quarter) of 2013.
Existing solution like heuristic based, visual similarity based take features from the
web page content so they take a lot of time to take decision. The phishing URL
classification scheme based only on investigative the suspicious URL and speed up
the running time of system. Therefore, in this paper, we proposed a machine
learning based phishing detection system which uses the URL features and analysed
it using naive Bayesian and SVM classifiers. Moreover, it does not require any
information from the e-content of the suspicious web page.
The remainder of this paper is organized as follows. Section 2 describes the
background and state-of-art techniques, its advantages and limitations. Section 3
describes our proposed phishing detection system in details. Evaluation of the
proposed system with results is discussed in Sect. 4. Finally, Sect. 5 concludes our
paper and discusses the scope for future work.
2 Related Work
There have been several techniques given in the literature to detect phishing attack
in last few years. In this section, we present an overview of detection approaches
against phishing attacks. Phishing detection approaches are broadly classified
into two types: user education based techniques and software-based techniques.
PHISH-SAFE: URL Features-Based Phishing Detection System … 469
In this section, we will discuss our proposed phishing detection system which can
detect a phishing page before user inputs personal information. Total 32,951
phishing URLs are taken from phishtank.com to evaluate the performance of the
proposed system. Following features are used for the phishing detection:
• IP Address: A phisher uses the IP address in place of domain name to hide the
identity of a website.
• Sub Domain: Phishing sites contain more than two sub-domains in URL. Each
domain is separated by dot (.). If any URL contain three or more than three dots,
then the probability of the suspicious site is more. In our experiment, we found
that 12,904 sites contain three or more number of dots.
• URL contains “@” symbol: the presence of “@” symbol in the URL ignore
everything previous to it. In our dataset, out of 32,951 phishing URL, 569 sites
contain @ symbol.
• Number of dash (-) in URL: To looks like genuine URL, phisher adds some
prefix or suffix with the brand name with dash, e.g., www.amazon-india.com.
We found that 42.5% of phishing URLs contain “dash” symbol.
• Length of URL: To hide the domain name, phisher uses the long URL. In our
experiment, we found the average length of URL is 74. We found that 7406
phishing sites contain length between 14 and 40 characters. 10,466 phishing
sites are having length between 41 and 60 characters. 6602 phishing URL
contain length between 61 and 80 character and 8475 sites contain length
between 81 and 2205 characters.
• Suspicious words in URL: Phishing URLs contain suspicious words such as
token, confirm, security, PayPal, login, signin, bank, account, update, etc., to
gain the trust on website. We have taken these nine frequently occurred words in
phishing sites.
• Position of Top-Level Domain: This feature checks the position of top-level
domain at proper place in URL.
Example—http://xyz.paypal.com.accounts.765issapidll.xtmll.ebmdata.com.
• Embedded Domain in URL: It checks this by checking for the occurrence of
“//” in the URL.
• HTTPS Protocol: HTTPS protocol is used for security. Phishing does not start
with https while legitimate URL provides security. (In our phishing dataset,
only 388 phishing sites contain https protocol).
• Number of times http appears: In phishing websites, http protocol may appear
more than one time but in genuine site, it appear only one time.
• Domains count in URL: Phishing URL may contain more than one domain in
URL. Two or more domains is used to redirect address.
• DNS lookup: If the DNS record is not available then the website is phishing.
The life of phishing site is very short, therefore; this DNS information may not
be available after some time.
PHISH-SAFE: URL Features-Based Phishing Detection System … 471
• Inconsistent URL: If the domain name of suspicious web page is not matched
with the WHOIS database record, then the web page is considered as phishing.
• Age of Domain: If the age of website is less than 6 month, then chances of fake
web page are more.
Training and testing of the proposed system are performed using following
classifiers:
(a) Naïve Bayes: Naïve Bayes is the probabilistic classifier, based on Bayes’
theorem with “naive” independence supposition. This classifier, used in text
categorization, can be an earning-based variant of keyword filtering. The rules
for decision making are explained below:
0 Pm Pn n o 1
ðiÞ ðiÞ
i¼1
i
j¼1 1 x j ¼ kandy ¼ 1 þ1
;kjy¼1 ¼ p xj ¼ kjy ¼ 1 ¼ @ Pm
ðI Þ ¼ 1gn þ jV j
A ð1Þ
i¼1 1fy i
0 Pm Pn n o 1
ðiÞ ðiÞ
i
i¼1 j¼1 1 x j ¼ k and y ¼ 0 þ1
;kjy¼1 ¼ p xj ¼ kjy ¼ 0 ¼ @ Pm
ðI Þ ¼ 1gn þ jV j
A ð2Þ
i¼1 1fy i
Pm
1 yðiÞ ¼ 1
;y¼1 ¼ i¼1
ð3Þ
ðm Þ
;xjy¼1 estimates the probability that a particular feature in a phishing URL will be
the k-th word in the dictionary. ;xjy¼0 estimates the probability that a particular
feature in a legitimate URL will be the k-th word in the dictionary. ;y estimates the
probability that any particular URL will be a phishing URL. m is the number of
URLs in our training set. The entire dictionary contains V words or the entire URLs
are V in number. For training, ;xjy¼0 ; ;xjy¼1 ; ;y are calculated and for testing,
p(x|y = 1) p(y = 1) is compared to p(x|y = 0) p(y = 0). To avoid underflow error,
logarithms are used. An email is classified as spam or phishing according to the
following equation:
In this section, we will discuss the tools and datasets used for implementation and
experiments results. The phishing detection using machine learning is classification
problem where system learns using various features of phishing and legitimate
URLs. After learning the system takes decision automatically based on training. We
have recognized various features of phishing and legitimate URLs discussed in the
previous section. We have collected 32,951 phishing URLs, taken from PhishTank
[9] and 2500 legitimate URLs taken from various sources.
Dataset Used: The dataset for phishing URLs is downloaded from PhishTank.
On 20th March 2015, a set of 32,951 phishing URLs were downloaded from
PhishTank. The datasets for non-phishing URLs are downloaded from Yahoo
Directory by using LinkKlipper from Chrome and DMOZ open directory.
Experiment Results: The feature extraction algorithm is implemented in Java and
the features of the URLs are stored in rows of a Sparse Matrix. A set of 15,000
training data (14,000 phishing URLs and 1000 non-phishing URLs) produced an
accuracy of 76.04%. A set of 25,000 training URLs (23,000 phishing URLs and
2000 non-phishing URLs) produced an accuracy of 91.28%. Phishing URL
detection using Naïve Bayes and SVM classifiers produced the results shown in
Table 1. From Table 1, it is found that when the size of the training set increases,
SVM performs better than Naïve Bayes classifier to detect phishing URL. Figure 2
shows the features contain by phishing URLs.
PHISH-SAFE: URL Features-Based Phishing Detection System … 473
This paper presented our proposed phishing detection system based on machine
learning. We have used 14 different features that distinguish phishing websites from
legitimate websites. Our experiment results show more than 90% accuracy in
detecting phishing websites using SVM classifier. In future, more features can be
added to improve the accuracy of the proposed phishing detection system.
Furthermore, other machine learning techniques can be used to increase the effi-
ciency of the proposed system.
References
Author Biographies
Ankit kumar Jain is presently working as Assistant Professor in National Institute of Technology,
Kurukshetra, India. He received Master of technology from Indian Institute of Information
Technology Allahabad (IIIT) India. Currently, he is pursuing PhD in cyber security from National
Institute of Technology, Kurukshetra. His general research interest is in the area of Information
and Cyber security, Phishing Website Detection, Web security, Mobile Security, Online Social
Network and Machine Learning. He has published many papers in reputed journals and
conferences.
474 A. K. Jain and B. B. Gupta
B. B. Gupta received Ph.D. degree from Indian Institute of Technology Roorkee, India in the area
of Information and Cyber Security. He published more than 100 research papers (including 02
books and 14 book chapters) in International Journals and Conferences of high repute including
IEEE, Elsevier, ACM, Springer, Wiley, Taylor & Francis, Inderscience, etc. His biography was
selected and published in the 30th Edition of Marquis Who’s Who in the World, 2012. Dr. Gupta
also received Young Faculty research fellowship award from Ministry of Electronics and
Information Technology, government of India in 2017. He is serving as associate editor of IEEE
Access and Executive editor of IJITCA, Inderscience, respectively. He is also serving as guest
editor of various reputed Journals. He was also visiting researcher with Yamaguchi University,
Japan in January 2015. At present, Dr. Gupta is working as Assistant Professor in the Department
of Computer Engineering, National Institute of Technology Kurukshetra India.
Semantic Security for Sharing
Computing Knowledge/Information
Abstract Due to the wide deployment of Internet and information technology for
storage and processing of data, the ever-growing privacy concern is the major
obstacle for information sharing. In the present digital scenario, the information
security is of prime concern. With hundreds of terabytes or even Petabytes of data/
information floating over around, it is important to have the access to the private
sensitive data only to authorized users. The explosive increase in the amount of
data/information leads to the growth of data mining techniques, a significant
resource for information security. The data mining is the extrication of relevant
patterns/ knowledge of information from bulk of data. It provides the variety of
applicable techniques, in accordance with the different security issues aroused, to
achieve a desired level of privacy. This paper provides a wide survey of the
emerging issues in the security field and various privacy-preserving techniques
PPDM techniques that can be used to mitigate the increasing security risks and
threats. It also centers on analyzing the problem of computation on private infor-
mation developing new concepts and techniques to deal with emerging privacy
issues in various contexts security of information while sharing and exchange using
Differential Privacy. Finally presents the challenges and techniques for differential
privacy as a trusted path to achieve privacy and discuss some of the theoretical and
practical challenges for future work in this area.
M. Narwaria (&)
School of Computer Science & Engineering, Galgotias University, Greater Noida, India
e-mail: [email protected]
S. Mishra
Department of Computer Applications, BSSS, Bhopal, India
e-mail: [email protected]
1 Introduction
Data mining is to extract required data from high dimensional databases. The big
databases are introduced by data mining application that are spread over the
business applications, which helps in predicting future trends, analyzing the data to
implement proactive decisions. The widespread availability of digital data in the
age of information, data analysis which goals at efficient and accurate discovery of
pattern and securing the private data at the same time, is the crucial task to be
performed. Differential privacy is one of the important techniques for releasing
statistical data without compromising the individual’s privacy. It reveals the queries
output from database at maximum accuracy while minimizing the chances of
identifying the individual private data.
Differential privacy has acknowledged much attention over some of the previously
used privacy algorithms, especially concentrates on the interactive setup for mini-
mizing the additive noise magnitude and checking the level of feasibility of dif-
ferentially private technique.
Barak et al. [1] discuss the technique which ensures that the non-negative
marginal count value and their sum are consistent for the problem of marginals of a
contingency table. Xiao et al. [1–38] and others address privelet, a wavelet
transformation-based approach for reducing the noise magnitude in the released
data to ensure differential privacy for multidimensional matrix. Hayes et al. [1–15]
specify a method one-dimensional dataset-based differentially private histograms.
Rastogi et al.{} design the mechanism for data perturbation which follows differ-
ential privacy. Machanavajjhala et al.{} proposed the technique for synthetic data
generation. {LAP 1}.
The concept of differential privacy is first given by Dwork et al. [22] which is
extended to the precise literature with McSherry. Micheal Schroeder coined the
term “differential privacy”. Dwork and Naor [24] then formulate the impossibility
of the semantic security. Composition and group privacy for(€,0)–differentially
private method is given by Dwork et al. [22]. Composition for(€,d) –differential
privacy was addressed by Dwork et al. [20] and then by Dwork and Lei [21].
Mironov proposed a mitigation against the vulnerability of the DP to inappropriate
implementation of real numbers.
Semantic Security for Sharing Computing Knowledge/Information 477
3 Differential Privacy
This is aimed at that there must be no difference in the response to the query
containing any particular individual or not. That is any adversary should unable to
frame or learn anything about the individual by querying the database. The two
ways for collecting and publishing data in the sanitized form are:
• Interactive
• Noninteractive
In the interactive approach, the data miner provides an interface to access the
dataset. The data miner queries the database holder using some private mechanism.
This is also referred as privacy-preserving distributed data mining (PPDDM).
In case of noninteractive setup, the data collector brings out an anonymized
version of data for analysis, which is collected as a result of applying sanitization
algorithm like permutation, subsampling and aggregation, identifier removal, etc. In
this approach, the horizontally partitioned data from different sources is securely
integrated without revealing the sensitive information. This approach gives more
flexibility than the former one.
Differential Privacy proposed by Dwork [19] is cryptographically motivated. It
ensures that the attacker cannot gain information about any data item in database by
simply querying the database. The approach imposes confidentiality by giving
perturbed query responses from database and provides a more robust privacy
guarantee. It is the framework which enables the analysis of privacy-sensitive
datasets and also ensures the privacy of individual-specific information. It is
flourishing as an area of research, including domains like computer security and
programming languages, statistics, databases, medical informatics, law, social sci-
ence. One of the research efforts is to reduce the error value that must be added to
query and analysis output keeping differential privacy. The other is to extend the
valuable existence of data for differentially private analyses.
PR ½MðDA Þ ¼ R
\ ¼ e€ ð1Þ
PR ½M ðDB Þ ¼ R
478 M. Narwaria and S. Mishra
where
PR is the probability of the perturbed query outcome of DA and DB,
M is the privacy granting function (Perturbation) on the query response from
database DA and DB,
R is the Perturbed query response from database DA and DB, and
e € is the exponential e epsilon value. € > 0, is public and specified by data
owner.
The technique provides a stronger privacy guarantee with the lower value of €.
Typically, the differential privacy is achieved by calibrating magnitude of noise to
the response of the query according to the sensitivity of the function. Where the
sensitivity is the maximum variation in the value due to addition or removal of a
single row.
Sensitivity:
Differential privacy provides protection from arbitrary risk. It also neutralizes the
linkage attacks automatically by including all the operations or information over
temporal dataset. Differential privacy technique has a measure of privacy loss and
also permits the analysis and its control incurred by groups. It compares a number
of techniques to find which technique provides better accuracy and privacy. The
behavioral properties of the differential privacy method under the composition
make it more complex than the other. Differential privacy is resistant to the
post-processing that is the analyst cannot computes the outcome of the function
from the differentially private algorithm and cannot increase the loss of privacy
without additional knowledge about the private database.
5 Research Methodology
See Table 1.
Table 1 Comparision of three privacy model in terms of data format and data size
K-anonymization De-identification Differential
privacy
Data format Structured Y N Y
Semi-structured N Y N
Un-structured N Y N
Data size Single record N Y N
required Dataset Y Y Y
Double dataset N N Y
480 M. Narwaria and S. Mishra
6 Conclusions
The study identifies the up-growing and promising areas of research where data
mining can be applied to accomplish protection of the information. The survey
literature also discusses the hottest trends and direction of research and relevance of
data mining in the safekeeping of information to improve the effectiveness of
privacy preserving/anonymization of data. The main aim of this study is to incor-
porate the recent DM techniques to remove the curbing in the privacy preservation
methodology. The adoption of data mining to secure the information which is the
vital resource is helpful to improve the performance in privacy preservation.
References
17. Dinur I, Nissim K (2003) Revealing information while preserving privacy. In: Proceedings of
the association for computing machinery SIGACTSIGMOD-SIGART symposium on
principles of database systems, pp 202–210
18. Duchi JC, Jordan MI, Wainwright MJ (2013) Local privacy and statistical minimax rates.
arXiv preprint arXiv:1302.3203
19. Dwork C (2006) Differential privacy. In: Proceedings of the international colloquium on
automata, languages and programming (ICALP), vol 2, pp 1–12
20. Dwork C, Kenthapadi K, McSherry F, Mironov I, Naor M (2006) Our data, ourselves: privacy
via distributed noise generation. In: EUROCRYPT, pp 486–503
21. Dwork C, Lei J (2009) Differential privacy and robust statistics. In: Proceedings of the 2009
international association for computing machinery symposium on theory of computing (STOC)
22. Dwork C, McSherry F, Nissim K, Smith A (2006) Calibrating noise to sensitivity in private
data analysis. In: Theory of cryptography conference’06, pp 265–284
23. Dwork C, McSherry F, Talwar K (2007) The price of privacy and the limits of lp decoding.
In: Proceedings of the association for computing machinery symposium on theory of
computing, pp 85–94
24. Dwork C, Naor M (2010) On the difficulties of disclosure prevention in statistical databases or
the case for differential privacy. J Priv Confid 2(1):93–107
25. Dwork C, Naor M, Pitassi T, Rothblum GN (2010) Differential privacy under continual
observation. In: Proceedings of the association for computing machinery symposium on
theory of computing, Association for Computing Machinery, pp 715–724
26. Dwork C, Naor M, Pitassi T, Rothblum GN, Yekhanin S (2010) Pan-private streaming
algorithms. In: Proceedings of international conference on super computing
27. Dwork C, Naor M, Reingold O, Rothblum GN, Vadhan SP (2009) On the complexity of
differentially private data release: efficient algorithms and hardness results. In Symposium on
theory of computing ’09, pp 381–390
28. Dwork C, Naor M, Vadhan S (2012) The privacy of the analyst and the power of the state. In:
Foundations of computer science
29. Dwork C, Nikolov A, Talwar K (2014) Efficient algorithms for privately releasing marginals
via convex relaxations. In: Proceedings of the annual symposium on computational geometry
(SoCG)
30. Dwork C, Nissim K (2004) Privacy-preserving datamining on vertically partitioned databases.
In: Proceedings of cryptology 2004, vol 3152, pp 528–544
31. Dwork C, Rothblum GN, Vadhan SP (2010) Boostingand differential privacy. In:
Foundations of computer science, pp 51–60
32. Dwork C, Talwar K, Thakurta A, Zhang L (2014) Analyze gauss: optimal bounds for
privacy-preserving pca. In: Symposium on theory of computing
33. Fleischer L, Lyu Y-H (2012) Approximately optimal auctions for selling privacy when costs
are correlated with data. In: Association for computing machinery conference on electronic
commerce, pp 568–585
34. Ghosh A, Ligett K (2013) Privacy and coordination: computing on databases with
endogenous participation. In: Proceedings of the fourteenth ACM conference on electronic
commerce (EC), pp 543–560
35. Ghosh A, Roth A (2011) Selling privacy at auction. In: Association for computing machinery
conference on electronic commerce, pp 199–208
36. Groce A, Katz J, Yerukhimovich A (2011) Limits of computational differential privacy in the
client/server setting. In: Proceedings of the theory of cryptography conference
37. Gupta A, Hardt M, Roth A, Ullman J (2011) Privately releasing conjunctions and the
statistical query barrier. In: Symposium on theory of computing’11, pp 803–812
38. Gupta A, Roth A, Ullman J (2012) Iterative constructions and private data release. In: Theory
of cryptography conference, pp 339–356
39. Dwork C, Roth A (2014) the algorithmic foundations of differential privacy. Found Trends
Theor Comput Sci 9(3–4):211–407
Paradigmatic Approach to Cloud
Security: Challenges and Remedies
1 Introduction
The word cloud computing was once a catchphrase around the globe, but now it has
become a mainstream. Cloud offers a platform to the user to use the application,
save the data, and access it when needed. It can be defined as the separation of the
applications from the system. Although the cloud provides a lot of benefits such as
ease of usability, flexibility, and accessibility but the major challenges associated
with cloud is the issue of security. Security is a prime obligation for cloud to work
as a strong and viable solution [1]. A similar thought has been also shared by
researchers’ corporate people [2], government organization [3, 4] and academics
[5]. Virtualization helps cloud providers to take off the control from the client and
manage their data. The user although has a security from other user’s but the data is
completely under the control of the service provider whose authenticity cannot be
verified by the cloud users. This poses a great threat to the data which is of utmost
importance. This issue has impacted the model’s creditability and popularity. The
cloud providers have to deal with security that requires a lot of expenses and
resources. The cloud security can be attained by implementing the three main
aspects namely:
• Strong protection to data,
• Complete control over the data, and
• Investment control.
If the trio is achieved by any means, then the threat to the security in cloud can
be surpassed. The efforts of identifying the risks and vulnerabilities have been done
by ENISA (European Network and Information Security Agency) [3] and the Cloud
Security Alliance (CSA) [4]. The documents of the two organizations present a
surfeit of security issues, recommended solutions dealing with privacy of data to
infrastructural arrangement.
2 Review of Literature
the methods to deal with the vulnerabilities and threats related to cloud computing.
The various countermeasures discussed in the paper are of utmost importance.
Mahalle and Shahade [8] have highlighted that using a combination of two
algorithms AES and RSA, two different files can be shared securely from data
transmission point of view.
Ficco et al. [9] in their work highlighted the effect of intrusion detection attack
on cloud environment and also explained the method for its prevention. The paper
focused on dynamic structure of the cloud. The model proposed here uses col-
laborative intrusion detection and prevention technique which basically works on
distributed cloud where the attacks are detected externally as well as internally.
It is quite evident that previous researches only focused on individual aspects of
cloud computing environment form security point but none of them used collab-
orative approach which is essential and works for infrastructure layer of cloud
(IaaS). In this paper, authors reveal the importance of collaboration rather than
tackling individual security issues at a time.
3 Problem Conceptualization
3.1 Methodology
Confidentiality
Availability Integrity
Paradigmatic Approach to Cloud Security: Challenges and Remedies 487
Data breach
Data loss
Account hijacking
Threats
Insufficient due diligence
Insecure API’s
Malicious insiders
Denial of service
4 Proposed Work
For achieving greater security the following observations helps to achieve that to a
great extent. The security should be managed from outside to inside. Here authors
investigate security issues from both service providers and users point of view and
suggested a concentric circular approach as a security factor (Fig. 3).
Figure 3 states that the security of data on cloud depends on the features pro-
vided by the cloud provider. At the first level, the architecture of cloud should be
considered that encompasses security of the network, virtualization, and interfaces.
The second aspect to be considered is related to the agreement between the cloud
user and cloud service provider. It encompasses the services provided to the user
and services to be provided by the provider. The third level to be considered is
related to privacy of data. This is considered to be the most crucial layer from user’s
488 R. Majumdar et al.
Agreement
Architecture
4.1 Architecture
The outermost circle named architecture should deal with security of network,
interfaces used, and virtualization. So the basic responsibility of this layer is to
group security, interface, and virtualization. The major issue concerning network
communication is the infrastructure of cloud. The solution provided should extend
customer’s present network structure [11]. It should utilize the existing local
security measures of the customer and enhance them to the network of cloud [12].
Following considerations should be made regarding architecture
• Security of data in transit: The channel used for transmitting the data should be
protected against spoofing, man-in-middle attack, sniffing, etc.
• Use of Firewall: As the firewalls are used at a small scale to provide security, it
should be extended to prevent denial of service attacks, and detect peripheral
security evaluation procedure.
• Security agreement: The protocols and technologies used at various levels in
cloud should be configured properly without hampering the privacy of data and
performance.
The proposed solution should also deal with issues related to the customer
interface, service provider interface, and the cloud interface.
Paradigmatic Approach to Cloud Security: Challenges and Remedies 489
4.2 Agreement
It deals with the norms and the requirements to provide service and availability of
data. It also deals with the kind of services provided and audit to be done by
customer, third party, and service providers regarding the cloud.
4.3 Privacy
It deals with security of data and the legal issues associated with it. The security of
data is provided by the technique of cryptography is used to encrypt the data [13].
A check should also be kept on data redundancy to ensure integrity and availability
of data. Legal issues include the location at which the data of the user reside, the
management of how the hardware is shared, and the privileges that are provided to
the user. The following diagram illustrates the proposed remedies in detail with
stepwise explanation.
Step 1:
Complete data about the client should be collected and what data he is willing to
store on cloud.
Step 2: Architecture
1. The data should be collected regarding which deployment or delivery model he
is willing to use.
2. On which server is the client hosting the application (on his own private server
or a third party server).
3. A check should be made by cloud provider for the server on which the data
needs to be stored.
4. Good authorization technique must be used.
5. A boundary for each user’s data should be designed and checked against
intrusion of data by unauthorized user.
6. The firewall should be checked to prevent DoS attack and evaluate peripheral
security
Step 3: Agreement
Safe and good techniques must be used for identity management.
Some security measures should be designed for virtualization manager.
An audit should be done by service provider, customer, and third party.
490 R. Majumdar et al.
Step 4: Privacy
Data should be encrypted while being stored.
Data that is flowing over the network should also be encrypted using network traffic
encryption technique such as secure socket layer and transport layer security.
When the data is being stored, it should be replicated for backup.
A proper strategy should be planned for business continuity and disaster recovery.
Data transactions should be safe and data integrity should be maintained.
Routing of data should be monitored.
The framework that has been proposed in our work provides a trade-off to the
existing approaches. The proposed work consists of the layered structure or shell
structure. The security check is levied on all the layers which apprehend the security
of the cloud architecture. Since the check is done layer after layer and no direct
access to the innermost layer, the vulnerability associated with the cloud security is
reduced to a great extent. The main objective of the security mechanism is to
provide ample measures to protect the stored data, data in transit, modified
encryption techniques are all enforced in our work. All the new features that have
been incorporated and realized in this work make it more secured measure as
compared to the existing approaches. As data is the king, so handling sensitive or
business-critical facts outside the organization will certainly lead risk because any
subcontracted service evades an organization’s in-house security panels easily. This
work investigates security in terms of risk and with cloud; one possible approach
will be to have a compatible control with preestablished dedicated service with
service providers. Organizations should try to learn about provider’s position for
greater control so that, it may not know exactly where its data resides or have any
ability to influence changes to the location of data. Most providers store data in a
shared environment, introduces security risk. No one security method will solve all
these data protection problems so it is important to consider multiple layers of
defense.
When adopting cloud services, among other critical factors we work on the
basics of where my data is, and how to handle new security threats?
Nowadays, security is no longer a source of worry. It has simply become another
reflection of hazard controlling policies and procedures.
Paradigmatic Approach to Cloud Security: Challenges and Remedies 491
6 Conclusion
References
Author Biography
A. B. Nimbalkar
Abstract This paper gives the survey of digital signatures, which are based on two
hard problems: (1) factorization (FAC) and (2) discrete logarithms (DL). In 1994,
L. Harn developed digital signature which is based on this two hard problems.
Z. Shao showed the drawback of L. Harn’s scheme and proposed his new scheme.
There are various schemes which are based on these two hard problems, because
attacker cannot solve both the problems simultaneously.
1 Introduction
A. B. Nimbalkar (&)
A.M. College, Pune, Maharashtra, India
e-mail: [email protected]
DL, ElGamal [3] invented new signature in 1985, which is used for message
confidentiality and digital signature schemes In 1985 elliptic curve cryptography
was invented which uses points of elliptic curve as group.
As asymmetric cryptosystem solve the repudiation problem, it is used to design
the digital signature schemes. A digital signature is a mathematical scheme for
validating the authenticity of a digital document. The digital signature has three
main phases. First phase is key generation. It includes the choice of two large prime
numbers and generates two keys, public and private keys. The second phase is
digital signature generation in which the message, keys, and modular arithmetic are
used to form the signature. The third is signature verification phase, where the
message is checked against the original message using the verification equation, if
equation satisfies then verifier believes that message is indeed an authenticated
message, else message is considered to be altered.
The digital signatures were developed using both the hard problems. In 1994, Harn
[4] designed new digital signature scheme based on FAC and DL. After this, there
were various schemes based on both the hard problems. Extensive research has been
done to show that without knowing private key, attacker can forge the signature.
2.1 L. Harn
There are several public key cryptosystems which are based on single cryptographic
assumption, such as factoring or discrete logarithms. In May 1994, Harn [4]
designed new scheme which is based on two hard problems, namely, FAC and DL.
To break this scheme attacker requires to solve Diffie–Hellman problem in a sub-
group of Zp as well as factoring a specific integer into product of two primes, both
of these problems are difficult. His scheme is based on two different cryptographic
assumptions for increasing the security and also maintaining the efficiency of
development. L. Harn maintains the computational time of the signature imple-
mentation that is maximum of RSA and ElGamal scheme.
The schemes which are based on single hard problem are not secure. It is known
that ElGamal scheme is less secure than RSA because computation of DL is easier
than the factorization. RSA schemes are efficient because it has message expansion
is one while in ElGamal message expansion is two. In July 1994, He and Kiesler [5]
designed two new signature schemes which are based on both the hard problems
that enhance the security. The message expansion in first version signature is two
but in second it is three.
The Digital Signature Schemes Based on Two Hard Problems … 495
In 1996 Lee and Hwang [6] showed that there is probability that the attacker can
forge the signatures of L. Harn schemes if he can solve DL modulo large prime
number. They showed that even if use of hash function there is possibility of attack.
The attack can be avoided if the condition that s′ is not allowed to equal to p′q′.
They proposed a modified L. Harn scheme which is based on both the hard
problems.
2.4 Z. Shao
Shao [7] designed two digital signature schemes in 1998. The security of these
signatures was equal to ElGamal and L. Harn signature schemes. There were some
drawbacks of ElGamal Scheme that are the size of public key is large, more
modulation was used, and every user uses his own public modulus. The substitution
attacks work on L. Harn signature scheme if one does not use one-way hash
function. On L. Harn scheme homomorphism attack gives private key x although
forging signature is not possible. If one wants to forge the signature then it is
necessary that one should able to find the cubic root modulo p − 1.
The Z. Shao scheme resists substitution and homomorphism attack. The effi-
ciency of Shao scheme is same as ElGamal and L. Harn scheme. In Z. Shao’s
scheme, only thing is that message expansion is three.
2.5 N. Y. Lee
The public key cryptographic algorithms are secure because there is no good
algorithm to solve the FAC and DL. Lee [8] in 1999 showed that the signatures
proposed by Z. Shao are not secure as he claimed. The security of Shao’s schemes
depends on only FAC and not DL. N. Y. Lee not proposed any scheme.
2.6 Z. Shao
Shao [9] in 2002 showed that there is forgery attack against Wei-Hua He’s sig-
nature scheme. If the attackers solve DL they can easily forge signature without
knowing private key of signer and this does not depend upon hardness of both FAC
and DL as claimed by Wei-Hua He. The task of designing new scheme based on
two hard problems was an open problem.
496 A. B. Nimbalkar
Wei [10] in 2004 tried to attack He–Kiesler’s scheme and showed that He–Kiesler
scheme does not resist his message attack. Based on this attack he designed new
scheme which resists such an attack and security based on two hard problems FAC
and DL. In 2007, Wei [11] improved Shao’s schemes [7] using quadratic residues
theory and proposed two new schemes based on two hard problems. Wei claimed
that his signatures resist Li_Xiao attack. The problem-solving quadratic equation is
equally hard as solving the FAC.
Zheng et al. [12] in 2008 showed that attacker can forge the Shimin Wei signature
schemes (2007). They showed that in Wei’s scheme, the universal forgery attack
can be possible and the two different messages (m) have same signature. Using
Wei’s scheme if we obtain the signature of one message then it is easy to obtain
signature of second message. This drawback can be remove by demanding
0 < m < p/2. They showed that universal forge attack can be done in less com-
putation than that of legal signer does. Also they showed that one can forge sig-
nature scheme for arbitrary message without knowing private keys. Hence the
security of Shimin Wei scheme is fails.
Ismail et al. [13] designed a new scheme in 2008, which provides better security by
using hash function. The main aim of designing the scheme was to increase the
security using FAC and DL. They showed five different attacks that were resisting
their signature. But time complexity in verification phase of signature was
increased.
In the year 2012, Swati Verma modified the Wei [10] scheme. The security of their
scheme depends primarily on two things: Use of one-way hash functions and the
intractability of solution to both DL and FAC simultaneously. They claimed that
this is more secure than earlier Wei [10] scheme.
3 Table of Comparison of Digital Signature Based on FAC and DL
Scheme Signature Secret key Public key Signing equation Verification
Elgamal [3] {m(r, s)} x is < p y =≺x mod p m = xr + ks mod (p − 1) m yr rs mod p
1 1
Harn Scheme [4] {m(k, r, s)} x is 1 x p − 1 y =≺x mod p m1 − ks1 + X1r mod pA − 1 s T
m
A ¼ r yA mod pA
2 2 2 2 2
He and Kiesleirs [5] {m(r, s, c)} x1 is 1 x1 n y ¼ gx mod p m = xr + ts (mod p − 1) gm ¼ yr r s g2rx mod p
x ¼ x1 2 mod (p − 1)
2 2 2
Shao Scheme [7] {m(k, r, s)} x = 1 < x < (p1q1/2) þ x2 þ x2 Þ þ k2
y ¼ gx mod p x1 s þ xr ¼ mt1 þ ktðmod p1 q1 Þ yðs ¼ rm g4ðmksxÞ mod p
ðx þ x1 Þ2 1 1 1 1 2 f 2 ðr r ; mÞ
Shao [9] {m(r1, r2, s)} R = p1q1 y¼g mod p ðx þ x Þ ¼ sðt þ t Þ þ f ðr1 ; r2 ; mÞ ðt þ t Þ mod R y r1s r2 1 2
x is god ((x + x−1)2, R) = 1 2sf
g ðr1 ; r2 ; mÞ mod p
2 4 2 2 2
Wei’s Scheme [11] {m(r1, r2, s)} x is 1 < x < n y ¼ gx mod p mt−1 = xr1 +ts2 (mod (p − 1)) r1s r2m ¼ yr g2ms
d c
Ismail et al. [13] {m(k, R, s)} x is 0 < x < n y ¼ gx mod p gs yhðmÞ K R Rk ðmod pÞ
The Digital Signature Schemes Based on Two Hard Problems …
4 Conclusion
The digital signatures based on single hard problem either Factorization or Discrete
Logarithm may not secure in future because computation can be possible. L. Harn
and He-Keisler gives the idea that if we combine these two problem then security
become more. Then there are many signature based on this two hard problem. Some
of signature can be forge without solving the hard problem. We make the com-
parative analysis of all digital signature schemes which are based on two hard
problems like Factorization and Discrete Logarithm. As security increases the Time
complexity also increases. The RSA signature has message expansion 1. ElGamal
has message expansion 2 and L. Harn has message expansion 3, as he uses both
hard problems for enhancing the security.
References
1. Diffie W, Hellman ME (1976) New directions in cryptography. IEEE Trans Inf Theory
22:644–654
2. Rivest RL, Shamir A, Adleman L (1978) A method for obtaining digital signatures and public
key cryptosystems. Commun ACM 21:120–126
3. ElGamal T (1985) A public key cryptosystem and a signature scheme based on discrete
logarithms. IEEE Trans Inf Theory IT-31(4):469–472
4. Harn L (1994) Public-key cryptosystem design based on factoring and discrete logarithms.
IEE Proc Comput Digital Techn 141:193–195
5. He J, Kiesler T (1994) Enhancing the security of ElGamal’s signature schemes. IEE Proc
Comput Digital Technol 141:249–252
6. Lee NY, Hwang T (1996) Modified Harn signature scheme based on factoring and discrete
logarithms. IEE Proc Comput Digital Tech 143:196–198
7. Shao Scheme Z (1998) Signature scheme based on factoring and discrete logarithms. IEE
Proc Comput Digital Tech 145(1)
8. Lee NY (1999) Security of Shao’s signature schemes based on factoring and discrete
logarithms. IEE Proc Control Theory Appl 146(2)
9. Shao Scheme Z (2002) Digital signature scheme based on factoring and discrete logarithms.
Electr Lett 38(24), 21 Nov 2002 (Online No: 20021093)
10. Wei S (2004) A new digital signature scheme based on factoring and discrete logarithms.
Progr Crypt Int J Ser Eng Comput Sci 769:107–111
11. Wei S (2007) Digital signature scheme based on two hard problems. Int J Comput Sci Netw
Secur 7(12)
12. Zheng J, Shao Z, Huang S, Yu T (2008) Security of two signature schemes based on two hard
problems. In: Proceedings of the 11th IEEE international conference on communication
technology, pp 745–748
13. Ismail ES, Thate NMF, Ahmad RR (2008) A new digital signature scheme based on integer
factorization and discrete logarithm. J Math Stat 4(4):222–225. ISSN 1549-3644
Gaussian Tendencies in Data Flow
in Communication Links
Abstract We have modeled data flow in communication link using random motion
of a particle, which results in a Gaussian pattern of traffic flow over a period of time.
The varying degrees of spectral deviation present a coherent model of data flow for
wired links. We have considered multiple link systems and presented an
n-dimensional representation of traffic model using a Gaussian function governed
by n-parameters. The model opens new insights toward analyzing and predicting
bandwidth requirements in communication links and their prospective failure.
1 Introduction
that it takes into account variation of traffic density with time. Gaussian distribution
can take into account spatial variation in traffic density across various locations
along a network.
Several economic and technological decisions are driven by model of data flow.
Sometimes, the data flow shows an exponential increase over a period of time
leading to extensive development and increase in channels. A drop in the
requirement has negative consequences on communication industry particularly
when the related investment has been made. This happened in the context of
telecom crisis of 2002 when it was assumed that the exponential rise in traffic flow
would continue for several years [4]. A correct data flow model might have averted
the crisis.
Random processes show random motion. Data flow between communication links
is random which implies that we can model it in terms of random motion. We
consider two data nodes A and B connected by a wire with a sensor S. The sensor
S generates a value 1 when data is transferred between A and B and 0 if there is no
data flow every second. We consider the density of 1 s and 0 s between A and
B. On carrying out a brief computational simulation of the process, we get a
Brownian motion as shown in Fig. 1, which shows a Gaussian distribution.
Random walk within a set of fixed constraints shows Gaussian distribution which is
given by the following equation [5]:
1 ðxlÞ2
pðxÞ ¼ pffiffiffiffiffiffi e 2r2 ð1Þ
r 2p
Find the correlation between time domain and frequency domain shown in
Fig. 4. This also shows Gaussian nature.
Gaussian distribution can also be spatial in nature. The density of traffic along a
route can vary in a Gaussian manner at a particular instant of time. This is evident in
general traffic network when the density of vehicles in high at the crossroads and is
sparse at distances away from the main traffic junctions. Figure 5 shows a simu-
lation of traffic density along a network line. The same graph shows the Fourier
Transform.
The peaks of Gaussian distribution are taken and plotted in Fig. 6 which shows a
set of Gaussian envelopes.
Figure 7 shows Traffic density along a network with time in Network Traffic of
Tata Communications Ltd., which shows a Gaussian distribution. The two sets of
data points are plotted and the envelope of the traffic density is similar to the graph
shown in Fig. 6 in distribution.
The Gaussian property is of more prominent nature. Although the empirical data
shows temporal distribution, the spatial distribution should have a similar profile.
504 R. P. Ojha et al.
6 Conclusion
We have modeled data flow using random motion and Gaussian curve, which
reflects the detailed physical behavior of traffic in a communication link. We
considered various drivers which influence data flow and change the topography of
the Gaussian curve. The work started with computational models and finally, the
results were verified through empirical data.
The model can help in making a more accurate forecast about requirements of
data in communication link and help in failure analysis. The model can also be used
in exploiting low-density points along the network nodes. There are times when the
spectral deviation starts decreasing and the Gaussian curve for a certain link starts
taking the form of an impulse function. When the rate of change of spectral
Gaussian Tendencies in Data Flow in Communication Links 505
deviation starts varying at a rate beyond a cut off value, the prospects of link
breakdown or traffic jamming would go up. Thus, a Gaussian angle of traffic
density can open new dimensions in traffic analysis. In future, the failure analysis
helps in the study of reliability of the system.
References
Author Biographies
Goutam Sanyal is designated as a Professor and Dean in Department of Computer Science and
Engineering. His qualifications are Bachelor’s of Engineering, Master’s of Technology, PhD
(Engineering), FIE (India), and MIEEE. He has more than 150 journal and research papers. He has
a work experience of 29 years in teaching and research and along with PhD guidance. His areas of
interest are wireless sensor network, computer architecture, computer graphics, computer vision,
image processing, VLSI, mathematical modelling and simulation.