0% found this document useful (0 votes)
13 views47 pages

CC Unit-4 Part2

The document discusses secure distributed data storage in cloud computing. It describes how cloud storage works and some common approaches like SANs and NAS. However, in cloud computing the storage nodes may not be under a single authority and attackers could compromise nodes. The document then analyzes security methods used by Amazon AWS, Microsoft Azure, and Google cloud services and identifies vulnerabilities. Specifically, it discusses how data integrity cannot be fully guaranteed between uploading and downloading from an untrusted cloud storage provider.

Uploaded by

vivek74543
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
13 views47 pages

CC Unit-4 Part2

The document discusses secure distributed data storage in cloud computing. It describes how cloud storage works and some common approaches like SANs and NAS. However, in cloud computing the storage nodes may not be under a single authority and attackers could compromise nodes. The document then analyzes security methods used by Amazon AWS, Microsoft Azure, and Google cloud services and identifies vulnerabilities. Specifically, it discusses how data integrity cannot be fully guaranteed between uploading and downloading from an untrusted cloud storage provider.

Uploaded by

vivek74543
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 47

SECURE DISTRIBUTED DATA

STORAGE IN CLOUD COMPUTING


○ A cloud actually is a collection of devices and
resources
● Connected through the Internet
○ One of the core services provided by cloud
computing is data storage
● How to create secure and reliable data storage
and access facilities
○ Over remote service providers in the cloud
● One of the necessary tasks to be addressed
before cloud computing is accepted
○ Data storage has been recognized as one of
the main concerns of IT
● The benefits of network-based applications have
led to the transition from server-attached
storage to distributed storage
○ A great quantity of efforts has been made
in the area of distributed storage security
● The research in cloud computing security is still
in its infancy
○ The unique issues of cloud computing
security have not been recognized
● Cloud computing security will not be much
different from existing security practices
● Well-managed using existing techniques
○ e.g., Digital signatures, encryption, firewalls,
and/or the isolation of virtual environments,
and so on
○ SSL (Secure Sockets Layer) is a
protocol
● Provides reliable secure communications on the
Internet
○ For things Web browsing, e-mail, instant messaging, and other
data transfers
○ The specific security requirements for
cloud computing have not been well-
defined within the community
● Warnings on the security threats in the cloud
computing model
● Potential users still wonder if the cloud is secure
○ At least two concerns when using the
cloud
● The users do not want to reveal their data to
the cloud service provider
○ The data could be sensitive information like medical records
● The users are unsure about the integrity of
the data they receive from the cloud.
● Within the cloud, more than conventional
security mechanisms will be required for data
security
○ Consider the recent research progress
and some results of secure distributed
data storage in cloud computing
○ The results of the migration from traditional
distributed data storage to the
cloud-computing-based data storage platform
○ A new vulnerability through analyzing three
current commercial cloud service platforms
○ The major difference brought by distributed
data storage in cloud computing
environment
○ Most designs of distributed storage take
the form of
●Either storage area networks (SANs) or
network-attached storage (NAS) on the LAN
level
○ e.g., The networks of an enterprise, a campus, or an
organization
● SANs are constructed on top of block-
addressed storage units
○ Connected through dedicated high-speed networks
● NAS is implemented by attaching
specialized file servers to a TCP/IP network
○ Providing a file-based interface to client machine
○ For SANs and NAS, the distributed storage
nodes are managed by the same authority
● The system administrator has control over
each node
● Essentially the security level of data is under
control
● The reliability of such systems is often
achieved by redundancy
● The storage security is highly dependent on
the security of the system against the attacks
and intrusion from outsiders
● The confidentiality and integrity of data are
mostly achieved using robust cryptographic
schemes
○ Global-scale collaboration over heterogeneous networks
under different authorities
● The specific data storage strategy is transparent to the
user
○ In a peer-to-peer (P2P) file sharing environment, or

the distributed storage in a cloud computing


environment
○ No approach to guarantee that data host nodes are under
robust security protection
● The activity of the medium owner is not controllable to
the data owner
○ Attackers can do whatever they wants to the data
stored in a storage node once the node is
compromised
○ Data storage services on the platform of cloud
computing are fundamentally provided by
applications/software based on the Internet
● User authentication, data confidentiality, and data
integrity can be solved through IPSec proxy
○ Using encryption and digital signature
● The key exchanging issues can be solved by SSL
proxy
● Applied to today’s cloud computing to secure the
data on the cloud
○ Also secure the communication of data to and from the cloud
● The service providers claim that their
services are secure
○ Three secure methods used in three
commercial cloud services and
discusses their vulnerabilities
○ Amazon’s Web Service
● Amazon provides Infrastructure as a
Service (IaaS) with different terms
○ e.g., Elastic Compute Cloud (EC2), SimpleDB,
Simple Storage Service (S3), and so on
○ Supposed to ensure the confidentiality, integrity,
and availability of the customers’ applications and
data
● The data processing methods adopted in
Amazon’s AWS
○ Used to transfer large amounts of data between the AWS
cloud and portable storage devices
● When the user wants to upload the data, stores
some parameters into an import metadata file
○ Called the manifest file
○ e.g., AccessKeyID, DeviceID, Destination, and so on
● Then signs the manifest file
○ E-mails the signed manifest file to Amazon
● Another metadata file is used by AWS
○ Named the signature file
○ To describe the cipher algorithm adopted to encrypt the job ID
and the bytes in the manifest file
 The log contains details about the data files that have been
uploaded
 Including the key names, number of bytes, and MD5

checksum values
 The downloading process is similar to the uploading process
 The user creates a manifest and signature file, e-mails the

manifest file, and ships the storage device attached with


signature file
 When Amazon receives these two files
 Validate the two files, copy the data into the storage device,

ship it back, and e-mail to the user with the status including
the MD5 checksum of the data
 Amazon claims that the maximum security is obtained via SSL
endpoints
○ Microsoft Windows Azure
● An Internet-scale cloud services platform hosted in
Microsoft data centers
● Provides an operating system and a set of developer
services
○ Can be used individually or together

● Provides scalable storage service


○ Three basic data items: blobs (up to 50 GB), tables,

and queues (<8k)


● Based on the blob, table, and queue structures
○ Microsoft promises to achieve confidentiality of the

users’ data
● The procedure for providing security for data
accessing to ensure that the data will not be
lost
● To use Windows Azure Storage service, a
user needs to create a storage account
○ Can be obtained from the Windows Azure portal web
interface
● The user will then receive a 256-bit secret
key
● Each time when the user wants to send the
data to or fetch the data from the cloud
○ The user has to use his secret key to create a HMAC
SHA256 signature for each individual request for
identification
● The user uses his signature to authenticate
request at server
○ The signature is passed with each request to
authenticate the user requests by verifying the HMAC
signature
● A REST request for a PUT/GET block
operation
● When the user wants to get the data, an
authorized data requests to Google Apps will
be first sent
○ As long as the data source is in the Google Apps domain to
the Google tunnel protocol servers
○ Forwards the request to the tunnel server
○ Storage services accepting a large amount of
data normally adopt strategies that help
make the shipment more convenient
● > 1 TB
● Just as the Amazon AWS does
○ Services that only accept a smaller data
amount allow the data to be uploaded or
downloaded via the Internet
● ≤ 50 GB
● Just as the Azure Storage Service does
○ To provide data integrity
○ The above procedure is secure for
each individual session
● The integrity of the data during the
transmission can be guaranteed by the
SSL protocol applied
○ From the perspective of cloud storage
services
● Data integrity depends on the security of
operations while in storage and the
security of the uploading and
downloading sessions
● The uploading session can only ensure
○ The data received by the cloud storage is the data
that the user uploaded
● The downloading session can simply guarantee
○ The data that the user retrieved is the data cloud storage
recorded
● This procedure applied on cloud storage services
cannot guarantee data integrity
○ Consider the following two scenarios
● Assume that Alice stores the company
financial data at a cloud storage service
provided by Eve
● Then Bob downloads the data from the cloud
○ Three important concerns in this simple
procedure
○ Confidentiality
○ Eve is considered as an untrustworthy third party
○ Alice and Bob do not want reveal the data to Eve
● Integrity
○ As the provider of the storage service, Eve has the
capability to play with the data in hand
○ How can Bob be confident that the data he fetched
from Eve are the same as what was sent by Alice?
○ Are there any measures to guarantee that the data
have not been tampered by Eve?
● Repudiation
○ If Bob finds that the data have been tampered with,
is there any evidence for him to demonstrate that it
is Eve who should be responsible for the fault?
○ Eve needs certain evidence to prove her innocence
○ The repudiation issue opens a door for
potentially blackmailers
● When the user is malicious
● Alice wants to blackmail Eve
● Eve is a cloud storage service provider
○ Claims that data integrity is one of their key features
● Alice stored some data in the cloud
○ Later she downloaded the data
● She reported that her data were incorrect
○ It is the fault of the storage provider
● Alice claims compensation for her so-called loss
● How can eve demonstrate her innocence?
○ Some ideas can bridge the missing link
Based on digital signatures and authentication

coding schemes
● Whether there is a Third Authority Certified
(TAC) by the user and provider
● Whether the user and provider are using the
Secret Key Sharing technique (SKS)
● Four solutions to bridge the missing link of data
integrity between the uploading and downloading
procedures
● Other digital signature technologies can also be
adopted to fix this vulnerability with different
approaches
○ Presents several technologies for
data security and privacy in cloud
computing
● Focusing on the unique issues of the
cloud data storage platform from a
few different perspectives
● Database Outsourcing and Query
Integrity Assurance
● Data Integrity in Untrustworthy
Storage
● Web-Application-Based Security
● Multimedia Data Security
○ Database outsourcing has become an
important component of cloud computing
● The cost of transmitting a terabyte of data over
long distances has decreased significantly in the
past decade.
● The total cost of data management is five to ten
times higher than the initial acquisition costs
● A growing interest in outsourcing database
management tasks to third parties
○ Can provide these tasks for a much lower cost due
to the economy of scale
● The benefits of reducing the costs for running
Database Management Systems (DBMS)
independently
○ Enabling enterprises to concentrate on their main businesses
○ The general architecture of a database
outsourcing environment with clients
● The database owner outsources its data management
tasks
● Clients send queries to the untrusted service provider
○ The data is preprocessed, encrypted, and stored at the service
provider
● For evaluating queries, a user rewrites a set of queries
against the data to queries against the encrypted
database
○ The outsourcing of databases to a third-party
service provider
● Two security concerns
○ Data privacy and query integrity
○ Data Privacy Protection
● A method to execute SQL queries over encrypted
databases
○ To process as much of a query as possible by the service
providers, without having to decrypt the data
○ Decryption and the remainder of the query processing are
performed at the client side
● An order-preserving encryption scheme for numeric
values
○ The transparent cloud provides flexible
utility of network-based resources
● The fear of loss of control on their data is one
of the major concerns that prevent end users
from migrating to cloud storage services
○ A potential risk that the storage infrastructure
providers become self-interested, untrustworthy, or
even malicious
○ Different motivations for a storage service
provider could become untrustworthy
● To cover the consequence of a mistake in
operation
● Or deny the vulnerability in the system after the data
have been stolen by an adversary
○ Two technologies to enable data owners to
verify the data integrity
● When the files are stored in the remote untrustworthy
storage services
○ Before cloud computing, several remote data
storage checking protocols have been
suggested
● In practice, a remote data possession checking
protocol has to satisfy the following five requirements
Requirement #1
○The verifier has to possess a complete copy of the data to be
checked
●Requirement #2
○ The protocol has to be very robust considering the
untrustworthy prover
○ Requirement #3
● The amount of information exchanged during the
verification operation should not lead to high
communication overhead
○ Requirement #4
● The protocol should be computationally efficient
○ Requirement #5
● It ought to be possible to run the verification an
unlimited number of times

○ An Enhanced Data Possession Checking
Protocol
● The above PDP-based protocol does not
satisfy Requirement #2 with 100%
probability
● An enhanced protocol has been proposed
based on the idea of the Diffie-Hellman
scheme
○ A specific method of securely exchanging cryptographic
keys over a public channel
● Satisfies all five requirements
● Computationally more efficient than the
PDP-based protocol
○ In cloud computing environments
● Resources are provided as a service over the
Internet in a dynamic, virtualized, and scalable
way
○ Users access business applications on-line from a
Web browser
○ Web security plays a more important role than ever
○ The Web site server is the first gate that
guards the vast cloud resources
● The cloud may operate continuously to process
millions of dollars’ worth of daily on-line
transactions
○ The impact of any Web security vulnerability will be
amplified at the level of the whole cloud
○ Web attack techniques are often
referred as the class of attack
● Attacker will employ those techniques to take
advantage of the security vulnerability
● The types of attack can be categorized in
○ Authentication, Authorization, Client-Side Attacks,
Command Execution, Information Disclosure, and
Logical Attacks
○ Authentication
● The process of verifying a claim that a subject
made to act on behalf of a given principal
● Authentication attacks target a Web site’s
method of validating the identity of a user,
service, or application
○ Including Brute Force, Insufficient Authentication, and
Weak Password Recovery Validation
● Brute Force attack employs an automated
process
○ To guess a person’s username and password by trial
and error
● In the Insufficient Authentication case
○ Some sensitive content or functionality are protected
by hiding the specific location in obscure string but
still remains accessible directly through a specific URL
○ The attacker could discover those URLs through a
Brute Force probing of files and directories
● Many Web sites provide password recovery
service
○ Automatically recover the user name or password to
the user if he can answer some questions defined as
part of the user registration process
○ If the recovery questions are either easily guessed or
can be skipped, this Web site is considered to be
Weak Password Recovery Validation
○ Authorization
● Used to verify if an authenticated subject can
perform a certain operation
○ Authentication must precede authorization
○ More and more multimedia contents are
being stored and delivered over many kinds
of devices, databases, and networks
● Multimedia Data Security plays an important role
in the data storage to protect multimedia data
○ How storage multimedia contents are delivered by
both different providers and users has attracted much
attentions and many applications
○ Protection from Unauthorized Replication
● Contents replication is required to generate
○ To keep multiple copies of certain multimedia
contents
○ e.g., Content distribution networks (CDNs) have been
used to manage content distribution to large numbers
of users
○ By keeping the replicas of the same contents on a
group of geographically distributed surrogates
● The replication can improve the system
performance
○ the unauthorized replication causes some problems
like contents copyright, waste of replication cost, and
extra control overheads
○ Protection from Unauthorized
Replacement
● The storage capacity is limited
○ A replacement process must be carried out when the
capacity exceeds its limit
○ A currently stored content must be removed from
the storage space to make space for the new
coming content
● How to decide which content should be removed
is very important
● If an unauthorized replacement happens
○ The content which the user doesn’t want to delete will
be removed resulting in an accident of the data loss
○ If the important content like system data is removed
by unauthorized replacement, the result will be more
serious
○ Protection from Unauthorized Pre-
fetching
● The Pre-fetching is widely deployed in
Multimedia Storage Network Systems
between server databases and end users’
storage disks
● If a content can be predicted to be requested
○ All the current commercial cloud service
providers adopt robust cipher algorithms
● For confidentiality of stored data
○ Also depend on network communication
security protocols to protect data in
transmission in the network
● e.g., SSL, IPSec, or others
○ Apply strong authentication and
authorization schemes in their cloud
domains
○ The requirement for a security cloud
computing is different from the traditional
security problems
● Encryption, digital signatures, network security,
firewalls, and the isolation of virtual environments
all are important for cloud computing security
● These alone will not make cloud computing reliable
for consumers

You might also like