IT436-week 3 - Physical Layer-Storage
IT436-week 3 - Physical Layer-Storage
Networking
Introduction
Acknowledgment: The presentations contains some figures and text from the following:
Cloud Computing Theory and Practice book by Dan C. Marinescu
Information Storage and Management, John Wiely & Sons, Inc.
Cloud Networking Understanding Cloud-based Data Center Networks by Gary Lee
Shaimaa M. Mohamed 1
Storage - Magnetic
Disks
▪ Circular platter constructed of nonmagnetic material, coated
with a magnetizable material.
▪ Data are recorded on and later retrieved from the disk via a
conducting coil named the head.
▪ The head is stationary while the platter rotates beneath it.
Storage - Magnetic
Disks
Storage - RAID
▪ Performance depends on the disk I/O operation.
▪ I/O operations depend on the computer system, the operating system,
and the nature of the I/O channel and disk controller hardware.
▪ The rate in improvement in secondary storage performance has been
considerably less than the rate for processors and main memory.
▪ This led to the development of arrays of disks that operate
independently and in parallel.
▪ With multiple disks, separate I/O requests can be handled in parallel.
▪ Furthermore, data can be organized and in which redundancy can be
added to improve reliability
Storage - RAID
▪ Improves storage performance by serving multiple write operations in
parallel.
▪ can provide protection against disk failure.
▪ Depends mainly on three techniques:
1. Stiping
2. Mirroring
3. Parity
Storage - RAID
Striping
▪ A technique to spread data across
multiple drives in order to use the
drives in parallel.
▪ Increase performance
▪ Each drive in a RAID group has a
predefined number of contiguously
addressable blocks called a “strip”.
▪ A set of aligned strips that span across
all the drives within the RAID group
is called a “stripe”.
▪ No data protection in case of disk
failure.
Storage - RAID
Mirroring
▪ A technique in which the same data is
stored simultaneously on two different
drives.
▪ Data protected against disk failure.
▪ Twice the amount of data being stored
is needed.
▪ Improves read performance because
read requests can be serviced by both
disks.
Storage - RAID
Parity
▪ Parity is a value derived by
performing a XOR on individual
strips of data and stored on a portion
of a RAID group.
▪ It enables the recreation of missing
data in case of a drive failure.
▪ Compared to mirroring, parity
implementation considerably reduces
the cost associated with data
protection.
▪ Parity is recalculated every time there
is a change in data, which may affect
the performance of the RAID array.
Storage - RAID Levels
▪ RAID levels are implementations using the three techniques explained.
▪ RAID 0 – that uses striping,
▪ RAID 1 – that uses mirroring,
▪ RAID 1+0 – which is a combination of RAID 1 and RAID 0,
▪ RAID 3, 5, and 6 – that use a combination of striping and parity.
Storage - RAID Levels
RAID 0
Storage - RAID Levels
RAID 1
Storage - RAID Levels
RAID 1 0
Storage - RAID Levels
RAID 3
▪ only a single redundant disk, no matter how large the disk array.
▪ Any I/O request will involve the parallel transfer of data from all of
the data disks
Storage - RAID Levels
RAID 5
▪ The distribution of parity strips across all drives avoids the potential
I/O bottleneck
Storage - RAID Levels
RAID 6
Depending on the type of the data access method used for a storage
system, the controller can either be classified:
▪ Block Level
▪ File Level
▪ Object Level
Storage
Accessing data from the Intelligent
Storage Systems
Storage
Block Level Access
Controller
Cache
Connectivity
Storage
Network
Ports Front-end
Controllers
Module 5: Block-based Storage System
Components of a Controller
Cache
Controller
Cache
Connectivity
Storage
Network
1. Read Request
2. Data sent to
compute system
Cache
4. Acknowledgment 3. Acknowledgment
Write-back cache
Cache
2. Acknowledgment 4. Acknowledgment
New Data
100%
HWM
LWM
Cache
Connectivity
Storage
Network
Back-end Ports
Controllers
Controller
Cache
Connectivity
Storage
Network
• A dedicated, high
performance file server with
storage (also known as
Network-attached Storage)
• Enables clients to share files
over an IP network
– Supports data sharing for
UNIX and Windows users
• Uses a specialized OS that is
optimized for file I/O
NAS Deployment Options
Storage
Network
LUN 1
LUN 1
Compute 2
LUN Expansion
MetaLUN
A method to expand LUNs that require additional
capacity or performance.
• Concatenated metaLUN
– Provides only additional capacity but no performance
– Expansion is quick as data is not restriped
• Striped metaLUN
– Provides capacity and performance
– Expansion is slow as data is restriped
Virtual Provisioning
10 TB
Thin Compute
LUN 0 System
Reported
Controller
Capacity
3 TB
Allocated Storage
Front End Back End (Storage Pool)
Thin
Compute 1 Cache LUN 0
Storage
Network
10 TB
Thin
LUN 1
Thin
LUN 1 Compute
System
Reported
Capacity
Compute 2 4 TB
Allocated
Expanding Thin LUNs
User
In-use capacity
capacity after
expansion
Thin LUN
User
capacity
Adding storage In-use before
drives to the Thin pool capacity expansion
storage pool rebalancing
Thin LUN
150 GB
Available
Capacity
1500 GB 1650 GB
or 800 GB or
1.5 TB 550 GB 1.65 TB
600 GB Allocated 500 GB Available
Allocated
Unused Capacity
Unused
400 GB
Allocated 500 GB Capacity Capacity
Unused Allocated
Unused
Capacity
Capacity 350 GB
350 GB 100 GB 50 GB 200 GB
100 GB 200 GB Actual data
Actual data Allocated Allocated Allocated
Data 50 GB Data Data
Thin Thin Thin Storage
LUN 1 LUN 2 LUN 3 Storage
LUN 1 LUN 2 LUN 3 System 2 TB
500 GB 550 GB 800 GB System 2 TB
Shaimaa M. Mohamed 64