Unit 5 CC
Unit 5 CC
UNIT 5
EVLUTION OF STORAGE TECHNOLOGY
Since the beginning of this century, advancements in data storage technology have
occurred at a breakneck speed, and the amount of information that is saved annually
has been steadily growing [233]: • 1986: 2.6 exabytes, which is equivalent to less
than one CD-ROM with 730 megabytes of storage space per person. • 1993 – 15.8 EB;
comparable to 4 CD-ROMs per person. • 2000 – 54.5 EB; comparable to 12 CD-ROMs per
person. • 2007 – 295 EB; equivalent to roughly 61 CD-ROMs per person
According to a study that was published in 2003 [354], the storage density of hard
disc drives (HDD) rose by four orders of magnitude between the years 1980 and 2003,
going from approximately 0.01 Gb/in2 to approximately 100 Gb/in2. At the same time
span, costs dropped by five orders of magnitude, reaching an all-time low of
approximately one cent per megabyte. It was anticipated that hard disc drive (HDD)
density will increase to 1,800 Gb/in2 by 2016, up from 744 Gb/in2 in 2011
Between the years 1990 and 2003, the density of DRAM, or dynamic random access
memory, increased from approximately 1 Gb/in2 to 100 Gb/in2. At the same time span,
the price of DRAM dropped from approximately $80 per megabyte to less than $1 per
megabyte. Using a 30 nm process Samsung was the first company to introduce LPDDR2
dram with four gigabytes in year 2010
Information can be stored not only using the charge of an electron, which is the
basis for solid state memories, but also using the spin of an electron, which is
another essential feature of an electron.
Although as the density of the storage devices has increased cost has been
decreased but the performance of this devices in accessing the data have been
increased. the performance of I/o when compared to the processor is not up to mark
which cause problem in accessing the applications like multi media engineering
software etc.
now a days data has been rapidly growing when compared to 1990 and 2000 coz evry
one are using mobile......
because of this it is unable to manage huge amount of data but data mining
algorithm came into existence........ and also cost of storage devices is also
more but using that in cloud computing would make it efficient.
FILE SYSTEM:CHATGPT
### **IBM GPFS (General Parallel File System) - Now Known as IBM Spectrum Scale**
IBM’s **General Parallel File System (GPFS)**, rebranded as **IBM Spectrum Scale**,
is a high-performance, scalable, and robust parallel file system designed for
managing large volumes of data. It was initially developed by IBM to support
supercomputing and large-scale enterprise applications, and it has evolved to meet
the needs of modern cloud and hybrid environments.
2. **Scalability**:
- GPFS can scale horizontally by adding more nodes and storage devices,
supporting **petabytes of data** and billions of files. It is designed to handle
both small and large files efficiently, making it suitable for diverse workloads,
from small data processing jobs to massive research simulations.
7. **Multi-Protocol Access**:
- GPFS supports multiple access protocols, including **POSIX**, **NFS**,
**SMB**, and **Object Storage API**, making it compatible with a wide range of
applications and environments. Users can access the file system using standard file
system interfaces, ensuring flexibility.
2. **Locking Mechanism**:
- GPFS uses a distributed locking mechanism to manage access to files. This
ensures that multiple clients can access the same file concurrently without data
corruption, while also preventing conflicts.
### **Conclusion**
IBM GPFS (Spectrum Scale) is a powerful, versatile file system designed to meet the
needs of high-performance, large-scale data environments. Its parallel
architecture, fault tolerance, and scalability make it a preferred choice for
industries requiring fast and reliable data access. As data workloads continue to
grow, the role of parallel file systems like GPFS becomes even more critical in
ensuring efficient data management and processing across on-premises, cloud, and
hybrid environments.
---
- *How it works: When you upload a file to cloud storage, that file is stored in a
specific **folder* on the server. When you want to access that file, you simply
navigate to the folder, find the file, and open it. The *path* of the file (for
example, *Documents/Resume.docx*) shows you where it is stored in the folder
hierarchy.
#### *Example*:
- *Advantages*:
- Simple to use, just like the way you organize files on your computer.
- Good for storing files like documents, images, videos, and presentations.
- Easy to share files with others by sharing folder links.
- *Limitations*:
- File systems are not ideal for large amounts of complex, structured data (like
customer records, inventory details, etc.).
- Searching through a large collection of files can be slower compared to using
databases.
- Difficult to establish relationships between different pieces of data.
---
Network File Systems (NFSs):
NFS is a protocol used for accessing files over a network, enabling systems to
share files between different machines. NFS allows a client system to access files
stored on a server as though they were local files.
Limitations:
NFS can become a bottleneck for performance due to scalability issues.
It can be unreliable since a failure in the NFS server can disrupt access to shared
files.
Storage Area Networks (SANs):
A *database* is a more advanced system used to store, manage, and organize large
amounts of *structured data. In a database, data is stored in **tables, which are
made up of **rows* and *columns. Each **row* represents a single record (like a
customer or an order), and each *column* represents an attribute of that record
(like the customer’s name, age, or address).
#### *Example*:
A simple *customer database table* might look like this:
- *Advantages*:
- Great for managing large amounts of data, especially if you need to organize it
in a structured way (e.g., customer data, transactions, inventory).
- Supports *relationships* between data (for example, connecting customers to
their orders).
- Allows for *efficient searches* and *complex queries* to find specific
information (e.g., “Find all customers older than 30”).
- Ensures data integrity and consistency with constraints and rules.
- *Limitations*:
- More complex to set up and use than a file system.
- Not ideal for unstructured data (like photos or videos).
- Requires more resources and management (e.g., database software, maintenance).
---