CC Mini Project
CC Mini Project
SUBMITTED BY
Vanshika Kotasthane
(T213035)
GUIDED BY
Asst. Apurva Sharma
Department of Computer
Engineering
Zeal Education Society’s
Acknowledgement
Abstract
The project will begin by setting up a Hadoop cluster with a master node and
multiple worker nodes. The cluster will be deployed on the LAN using
opensource technologies such as Apache Hadoop and Apache Spark. HDFS will
be used as the primary storage solution, providing a distributed file system for
storing large amounts of data across multiple nodes in the cluster.
Once the cluster is set up, the focus will shift to developing a cloud controller
that can handle basic operations such as file segmentation and uploading. The
cloud controller will be developed using open-source technologies such as
Python and Flask. The controller will use HDFS APIs to interact with the file
system and perform operations such as file segmentation and uploading.
The file segmentation process will involve dividing large files into smaller
segments or blocks, which can be distributed across the worker nodes in the
Hadoop cluster. This will help to optimise the performance of the system and
reduce the time taken for file transfers.
The uploading process will involve sending the segmented files to the cloud
controller, which will then upload them to the HDFS file system. The controller
will also keep track of the file locations and metadata, allowing users to easily
access and retrieve their files from the cloud.
Overall, this project will provide a practical and cost-effective solution for setting
up a private cloud for SaaS over an existing LAN in a laboratory. It will
demonstrate the use of open-source technologies such as Hadoop and Flask to
develop a cloud controller that can handle basic operations such as file
segmentation and uploading.
Introduction
Cloud computing has revolutionised the way we store and access data, making
it easier and more cost-effective to manage large amounts of data. Software as
a Service (SaaS) is a popular cloud computing model that allows users to access
software applications over the internet without the need for local installation.
In this project, we aim to set up a private cloud for SaaS over an existing Local
Area Network (LAN) in a laboratory. The cloud will be implemented using open-
source technologies, making it a cost-effective solution for small-scale
deployments.
The cloud will use the Hadoop Distributed File System (HDFS) as the primary
storage solution, providing a distributed file system for storing large amounts of
data across multiple nodes in the cluster. The Hadoop cluster will consist of a
master node and multiple worker nodes, which will be deployed on the LAN
using open-source technologies such as Apache Hadoop and Apache Spark.
To interact with the cloud, we will develop a cloud controller using open-source
technologies such as Python and Flask. The cloud controller will provide a
webbased interface for users to upload and access their files stored in the cloud.
The controller will also handle basic operations such as file segmentation and
uploading, making it easier for users to manage their files in the cloud.
Overall, this project will provide a practical and cost-effective solution for setting
up a private cloud for SaaS over an existing LAN in a laboratory. The project will
demonstrate the use of open-source technologies such as Hadoop and Flask to
develop a cloud controller that can handle basic operations such as file
segmentation and uploading.
Project
Setup your own cloud for Software as a Service (SaaS) over the existing LAN in
your laboratory. In this assignment you have to write your own code for a cloud
controller using open-source technologies to implement with HDFS.
Implementing the basic operations may be like to divide the file in
segments/blocks and upload/ download file on/from cloud in encrypted form.
Implementation screenshots:
Accept agreement
Click on Done
proceed further
Enter (Continue)
Conclusion
The development of a cloud controller using Python and Flask has allowed for
easy file segmentation and uploading, making it easier for users to manage their
files in the cloud. The HDFS file system has provided a reliable and scalable
storage solution for the cloud, allowing for easy expansion as data storage needs
grow.
Overall, this project has provided a practical and cost-effective solution for
setting up a private cloud for SaaS in a laboratory setting. It has demonstrated
the use of open-source technologies to develop a cloud controller that can
handle basic operations such as file segmentation and uploading, making it
easier for users to manage their files in the cloud. The project can be further
extended to include additional features such as data backup and recovery,
access control, and user management.
References