Dotnet: Towards Secure and Dependable Storage Services in Cloud Computing
Dotnet: Towards Secure and Dependable Storage Services in Cloud Computing
HEURISTICS BASED QUERY PROCESSING FOR LARGE RDF GRAPHS USING CLOUD COMPUTING
0emantic $eb is an emerging area to augment human reasoning. :arious technologies are being developed in this arena $hich have been standardi)ed by the World Wide Web 3onsortium *W;3+. <ne such standard is the =esource 7escription >rame$or% *=7>+. 0emantic $eb technologies can be utili)ed to build efficient and scalable systems for 3loud 3omputing. With the e'plosion of semantic $eb technologies, large =7> graphs are common place. 5his poses significant challenges for the storage and retrieval of =7> graphs. 3urrent frame$or%s do not scale for large =7> graphs and as a result do not address these challenges. &n this paper, $e describe a frame$or% that $e built using 4adoop to store and retrieve large numbers of =7> triples by e'ploiting the cloud computing paradigm. We describe a scheme to store =7> data in 4adoop 7istributed >ile 0ystem. ?ore than one 4adoop job *the smallest unit of e'ecution in 4adoop+ may be needed to ans$er a query because a single triple pattern in a query cannot simultaneously ta%e part in more than one join in a single 4adoop job. 5o determine the jobs, $e present an algorithm to generate query plan, $hose $orst case cost is bounded, based on a greedy approach to ans$er a 06A=12 6rotocol and =7> 1uery 2anguage *06A=12+ query. We use 4adoop@s ?ap=educe frame$or% to ans$er the queries. <ur results sho$ that $e can store large =7> graphs in 4adoop clusters built $ith cheap commodity class hard$are. >urthermore, $e sho$ that our frame$or% is scalable and efficient and can handle large amounts of =7> data, unli%e traditional approaches.
the user the cloud should provide a $ay for the user to chec% if the integrity of his data is maintained or is compromised. &n this paper $e provide a scheme $hich gives a proof of data integrity in the cloud $hich the customer can employ to chec% the correctness of his data in the cloud. 5his proof can be agreed upon by both the cloud and the customer and can be incorporated in the 0ervice level agreement *02A+. 5his scheme ensures that the storage at the client side is minimal $hich $ill be beneficial for thin clients.
EXPLOITING DYNAMIC RESOURCE ALLOCATION FOR EFFICIENT PARALLEL DATA PROCESSING IN THE CLOUD
&n recent years ad hoc parallel data processing has emerged to be one of the %iller applications for &nfrastructure-as-a-0ervice *&aa0+ clouds. ?ajor 3loud computing companies have started to integrate frame$or%s for parallel data processing in their product portfolio, ma%ing it easy for customers to access these services and to deploy their programs. 4o$ever, the processing frame$or%s $hich are currently used have been designed for static, homogeneous cluster setups and disregard the particular nature of a cloud. 3onsequently, the allocated compute resources may be inadequate for big parts of the submitted job and unnecessarily increase processing time and cost. &n this paper, $e discuss the opportunities and challenges for efficient parallel data processing in clouds and present our research project /ephele. /ephele is the first data processing frame$or% to e'plicitly e'ploit the dynamic resource allocation offered by today@s &aa0 clouds for both, tas% scheduling and e'ecution. 6articular tas%s of a processing job can be assigned to different types of virtual machines $hich are automatically instantiated and terminated during the job e'ecution. (ased on this ne$ frame$or%, $e perform e'tended evaluations of ?ap=educe-inspired processing jobs on an &aa0 cloud system and compare the results to the popular data processing frame$or% 4adoop.