0% found this document useful (0 votes)
113 views32 pages

Unit IV Backup - Lectures Notes

Uploaded by

SHAAFIA PARVEEN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
113 views32 pages

Unit IV Backup - Lectures Notes

Uploaded by

SHAAFIA PARVEEN
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 32

Unit No

BACKUP, ARCHIVE AND REPLICATION Lecture No 1


4

Introduction to Business Continuity, Information Availability and Measures, Backup and


Topic
Recovery.

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Define Business Continuity K1

LO2 Explain Information Availability and causes of Information K2


Unavailability.

LO3 Discuss the measures of Information Availability. K2

LO4 Differentiate Backup and Recovery. K2

Lecture notes

Introduction to Business Continuity


Business continuity (BC) is an integrated and enterprise wide process that includes all activities (internal
and external to IT) that a business must perform to mitigate the impact of planned and unplanned
downtime. BC entails preparing for, responding to, and recovering from a system outage that adversely
affects business operations. It involves proactive measures, such as business impact analysis and risk
assessments, data protection, and security, and reactive countermeasures, such as disaster recovery and
restart, to be invoked in the event of a failure. The goal of a business continuity solution is to ensure the
“information availability” required to conduct vital business operations.

Information Availability
Information availability (IA) refers to the ability of the infrastructure to function according to business
expectations during its specified time of operation.
Information availability ensures that people (employees, customers, suppliers, and partners) can access
information whenever they need it. Information availability can be defined with the help of reliability,
accessibility and timeliness.
Reliability: This reflects a component’s ability to function without failure, under stated conditions, for a
specified amount of time.
Accessibility: This is the state within which the required information is accessible at the right place, to
the right user. The period of time during which the system is in an accessible state is termed system
uptime; when it is not accessible it is termed system downtime.
Timeliness: Defines the exact moment or the time window (a particular time of the day, week, month,
and/or year as specified) during which information must be accessible. For example, if online access to an
application is required between 8:00 am and 10:00 pm each day, any disruptions to data availability
outside of this time slot are not considered to affect timeliness.
Causes of Information Unavailability
Various planned and unplanned incidents result in data unavailability. Planned outages include
installation/integration/maintenance of new hardware, software upgrades or patches, taking backups,
application and data restores, facility operations (renovation and construction), and refresh/migration of
the testing to the production environment. Unplanned outages include failure caused by database
corruption, component failure, and human errors.
Another type of incident that may cause data unavailability is natural or man-made disasters such as
flood, fire, earthquake, and contamination. As illustrated in Figure 11-1, the majority of outages are
planned. Planned outages are expected and scheduled, but still cause data to be unavailable. Statistically,
less than 1 percent is likely to be the result of an unforeseen disaster.

Measuring Information Availability


Information availability relies on the availability of the hardware and software components of a data
center. Failure of these components might disrupt information availability. A failure is the termination of
a component’s ability to perform a required function. The component’s ability can be restored by
performing an external corrective action, such as a manual reboot, a repair, or replacement of the failed
component(s). Repair involves restoring a component to a condition
that enables it to perform a required function within a specified time by using procedures and resources.
Proactive risk analysis performed as part of the BC planning process considers the component failure rate
and average repair time, which are measured by MTBF and MTTR:
■Mean Time Between Failure (MTBF): It is the average time available for a system or component to
perform its normal operations between failures.
■Mean Time To Repair (MTTR): It is the average time required to repair a failed component. While
calculating MTTR, it is assumed that the fault responsible for the failure is correctly identified and that
the required spares and personnel are available. Note that a fault is a physical defect at the component
level, which may result in data unavailability. MTTR includes the time required to do the following:
detect the fault, mobilize the maintenance team, diagnose the fault, obtain the spare parts, repair, test, and
resume normal operations.
IA is the fraction of a time period that a system is in a condition to perform its intended function upon
demand. It can be expressed in terms of system uptime and downtime and measured as the amount or
percentage of system uptime:
IA = system uptime / (system uptime + system downtime)
In terms of MTBF and MTTR, IA could also be expressed as
IA = MTBF / (MTBF + MTTR)

Consequences of Downtime
Data unavailability, or downtime, results in loss of productivity, loss of revenue, poor financial
performance, and damages to reputation. Loss of productivity reduces the output per unit of labor,
equipment, and capital. Loss of revenue includes direct loss, compensatory payments, future revenue
losses, billing losses, and investment losses. Poor financial performance affects revenue recognition,
cash flow, discounts, payment guarantees, credit rating, and stock price. Damages to reputation may result
in a loss of confidence or credibility with customers, suppliers, financial markets, banks, and business
partners.
Other possible consequences of downtime include the cost of additional equipment rental, overtime, and
extra shipping.
The business impact of downtime is the sum of all losses sustained as a result of a given disruption. An
important metric, average cost of downtime per hour, provides a key estimate in determining the
appropriate BC solutions.
It is calculated as follows:
Average cost of downtime per hour = average productivity loss per hour + average revenue loss per hour
Where:
Productivity loss per hour = (total salaries and benefits of all employees per week) / (average number of
working hours per week)
Average revenue loss per hour = (total revenue of an organization per week) / (average number of hours
per week that an organization is open for business)
The average downtime cost per hour may also include estimates of projected revenue loss due to other
consequences such as damaged reputations and the additional cost of repairing the system.

Backup and Recovery

A backup is a copy of production data, created and retained for the sole purpose of recovering deleted or
corrupted data. With growing business and regulatory demands for data storage, retention, and
availability, organizations are faced with the task of backing up an ever-increasing amount of data. This
task becomes more challenging as demand for consistent backup and quick restore of data increases
throughout the enterprise which may be spread over multiple sites. Moreover, organizations need to
accomplish backup at a lower cost
with minimum resources.

Backups are performed to serve three purposes: disaster recovery, operational backup, and archival.

Disaster Recovery
Backups can be performed to address disaster recovery needs. The backup copies are used for restoring
data at an alternate site when the primary site is incapacitated due to a disaster. Based on RPO and RTO
requirements, organizations use different backup strategies for disaster recovery. When a tape-based
backup method is used as a disaster recovery strategy, the backup tape media is shipped and stored at an
offsite location. These tapes can be recalled for restoration at the disaster recovery site. Organizations
with stringent RPO and RTO
requirements use remote replication technology to replicate data to a disaster recovery site. This allows
organizations to bring up production systems online in a relatively short period of time in the event of a
disaster.

Operational Backup
Data in the production environment changes with every business transaction and operation. Operational
backup is a backup of data at a point in time and is used to restore data in the event of data loss or logical
corruptions that may occur during routine processing. The majority of restore requests in most
organizations
fall in this category. For example, it is common for a user to accidentally delete an important e‑mail or for
a file to become corrupted, which can be restored from operational backup.
Operational backups are created for the active production information by using incremental or differential
backup techniques, detailed later in this chapter. An example of an operational backup is a backup
performed for a production database just before a bulk batch update. This ensures the availability of a
clean copy of
the production database if the batch update corrupts the production database.
Archival
Backups are also performed to address archival requirements. Although CAS has emerged as the primary
solution for archives, traditional backups are still used by small and medium enterprises for long-term
preservation of transaction records, e‑mail messages, and other business records required for regulatory
compliance.

Apart from addressing disaster recovery, archival, and operational requirements, backups serve as a
protection against data loss due to physical damage of a storage device, software failures, or virus attacks.
Backups can also be used to protect against accidents such as a deletion or intentional data destruction.

Backup Granularity
Most organizations use a combination of these three backup types to meet their backup and recovery
requirements
Full backup is a backup of the complete data on the production volumes at a certain point in time. A full
backup copy is created by copying the data on the production volumes to a secondary storage device.
Incremental backup copies the data that has changed since the last full or incremental backup, whichever
has occurred more recently. This is much faster (because the volume of data backed up is restricted to
changed data), but it takes longer to restore.
Cumulative (or differential) backup copies the data that has changed since the last full backup. This
method takes longer than incremental backup but is faster to restore.

Assessment questions to the lecture


Qn Bloom’s
Question Answer
No Knowledge Level

1. Which of the following is NOT a common threat c. K1


addressed by business continuity planning?
a) Natural disasters
b) Cyber-attacks
c) Employee training programs
d) Supply chain disruptions
2. What does the term "information availability" b. K1
refer to?
a) The amount of data stored in a system
b) The ease of access to information when
needed
c) The security measures implemented to protect
information
d) The process of sharing information with
others
3. What is the primary goal of implementing c. K1
backup and recovery procedures?
a) To prevent unauthorized access to information
b) To ensure the confidentiality of sensitive data
c) To minimize data loss and maintain
information availability
d) To improve the performance of information
systems

Students have to prepare answers for the following questions at the end of the lecture

Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 What is Business Continuity? 2 4 K1

2 Discuss about Information Availability and the causes of


8 4 K2
Information Unavailability.

3 Differentiate Backup and Recovery. 2 4 K2

4 Explain the measures of Information Availability. 8 4 K2

Reference Book

Author(s) Title of the book Page numbers

G Somasundaram EMC Corporation, Information Storage and 251-253


Management, Wiley, India
Alok Shrivastava
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 2
No 4

Backup targets and methods-Backup Process and Architecture- Backup and Restore
Topic
Operations

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Discuss the backup methods and targets K1

LO2 Explain the Backup process and Architecture. K2

LO3 Explain Backup and Restore operations. K2

LO4 Differentiate Backup and Restore. K2

Lecture notes

Backup targets and access methods

Cloud storage providers, including Amazon Web Services (AWS) and Microsoft Azure, offer
Infrastructure as a Service (IaaS) models that let you create servers in the cloud as backup targets.
Most cloud backup services provide dedicated solutions that look like network drives to software and
users. While this option provides a high level of flexibility, it also comes with additional fees. You might
also incur additional costs needed to protect the data stored in these services.
Cloud customers can also employ a local storage resource, such as a network-attached storage (NAS)
device, to serve as a middleman for backups. This type of resource can store frequently used files and
then serve them through the faster local network.

Management and storage tiers

Cloud backup vendors usually provide management tools according to customer requirements in terms of
size and demand, security, and changing bandwidth conditions. In some scenarios, variable data retention
requirements are also factored in. This enables cloud providers to automatically drop all files or folders
that are older than the time specified by the administrator.
Storage tiers, like frequently-accessed storage or archival storage, are priced differently. Customers can
define policies and automate how data moves between tiers, to conserve costs for data that is less
frequently used or retained only for compliance purposes.

Backup Methods
Hot backup and cold backup are the two methods deployed for backup. They are based on the state of the
application when the backup is performed. In a hot backup, the application is up and running, with users
accessing their data during the backup process. In a cold backup, the application is not active during the
backup process.
The backup of online production data becomes more challenging because data is actively being used and
changed. An open file is locked by the operating system and is not copied during the backup process until
the user closes it. The backup application can back up open files by retrying the operation on files that
were opened earlier in the backup process. During the backup process, it may be possible that files
opened earlier will be closed and a retry will be successful.
The maximum number of retries can be configured depending on the backup application. However, this
method is not considered robust because in some environments certain files are always open.

Backup Process
A backup system uses client/server architecture with a backup server and multiple backup clients. The
backup server manages the backup operations and maintains the backup catalog, which contains
information about the backup process and backup metadata. The backup server depends on backup clients
to gather the data to be backed up. The backup clients can be local to the server or they can reside on
another server, presumably to back up the data visible to that server. The backup server receives backup
metadata from the backup clients to perform its activities.
Figure 12-4 illustrates the backup process. The storage node is responsible for writing data to the backup
device (in a backup environment, a storage node is a host that controls backup devices). Typically, the
storage node is integrated with the backup server and both are hosted on the same physical platform. A
backup device is attached directly to the storage node’s host platform. Some backup architecture refers to
the storage node as the media server because it connects to the storage device. Storage nodes play an
important role in backup planning because they can be used to consolidate backup servers.
The backup process is based on the policies defined on the backup server, such as the time of day or
completion of an event. The backup server then initiates the process by sending a request to a backup
client (backups can also be initiated by a client). This request instructs the backup client to send its
metadata
to the backup server, and the data to be backed up to the appropriate storage node. On receiving this
request, the backup client sends the metadata to the backup server. The backup server writes this metadata
on its metadata catalog.
The backup client also sends the data to the storage node, and the storage node writes the data to the
storage device. After all the data is backed up, the storage node closes the connection to the backup
device. The backup server writes backup completion status to the metadata catal

Backup software also provides extensive reporting capabilities based on the backup catalog and the log
files. These reports can include information such as the amount of data backed up, the number of
completed backups, the number of incomplete backups, and the types of errors that may have occurred.
Reports can be customized depending on the specific backup software used.
Backup and Restore Operations
When a backup process is initiated, significant network communication takes place between the different
components of a backup infrastructure. The backup server initiates the backup process for different clients
based on the backup schedule configured for them. For example, the backup process for a group of clients
may be scheduled to start at 3:00 am every day.
The backup server coordinates the backup process with all the components in a backup configuration (see
Figure 12-5). The backup server maintains the information about backup clients to be contacted and
storage nodes to be used in a backup operation. The backup server retrieves the backup-related
information
from the backup catalog and, based on this information, instructs the storage node to load the appropriate
backup media into the backup devices.
Simultaneously, it instructs the backup clients to start scanning the data, package it, and send it over the
network to the assigned storage node. The storage node, in turn, sends metadata to the backup server to
keep it updated about the media being used in the backup process. The backup server continuously
updates the backup catalog with this information.

Upon receiving a restore request, an administrator opens the restore application to view the list of clients
that have been backed up. While selecting the client for which a restore request has been made, the
administrator also needs to identify the client that will receive the restored data. Data can be restored on
the same client for whom the restore request has been made or on any other client.
The administrator then selects the data to be restored and the specified point in time to which the data has
to be restored based on the RPO. Note that because all of this information comes from the backup catalog,
the restore application must also communicate to the backup server.

Assessment questions to the lecture


Qn Bloom’s
Question Answer
No Knowledge Level

1. Which of the following is NOT a common backup c. K1


target?

a) External hard drive

b) Cloud storage

c) Main memory

d) Tape drive

2. Which component of the backup architecture is b. K1


responsible for managing and executing backup
operations?

a) Backup client

b) Backup server

c) Backup storage

d) Backup application

3 Which factor is crucial for determining the speed of c. K1


a restore operation?

a) Backup encryption

b) Compression ratio

c) Backup storage type

d) Backup size
Students have to prepare answers for the following questions at the end of the lecture
Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 List the backup methods. 2 2 K1

2 Differentiate Backup and Restore. 2 2 K1

3 Explain the Backup process and Architecture. 8 2 K2

4 Explain Backup and Restore operations. 8 2 K2

Reference Book

Author(s) Title of the book Page numbers

G Somasundaram EMC Corporation, Information Storage and 258-260


Management, Wiley, India
Alok Shrivastava
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 3
No 4

Topic Data deduplication- Techniques-File level and Block level deduplication

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Define Data Deduplication. K1

LO2 Describe the techniques to deduplicate data K2

LO3 Explain File level and Block level deduplication K2

LO4 Discuss the benefits and drawbacks of Deduplication K2

Lecture notes

Data Deduplication
Data deduplication is a process that eliminates redundant copies of data and reduces storage overhead.
Data deduplication techniques ensure that only one unique instance of data is retained on storage media,
such as disk, flash or tape. Redundant data blocks are replaced with a pointer to the unique data copy. In
that way, data deduplication closely aligns with incremental backup, which copies only the data that has
changed since the previous backup.
An example of data deduplication
A typical email system might contain 100 instances of the same 1 megabyte (MB) file attachment. If the
email platform is backed up or archived, all 100 instances are saved, requiring 100 MB of storage space.
With data deduplication, only one instance of the attachment is stored; each subsequent instance is
referenced back to the one saved copy.
In this example, a 100 MB storage demand drops to 1 MB.

Target vs. source deduplication


Data deduplication can occur at the source or target level.
Source-based dedupe removes redundant blocks before transmitting data to a backup target at the client or
server level. There is no additional hardware required. Deduplicating at the source reduces bandwidth and
storage use.
In target-based dedupe, backups are transmitted across a network to disk-based hardware in a remote
location. This deduplication increases costs, although it generally provides a performance advantage
compared to source dedupe, particularly for petabyte-scale data sets.

Techniques to deduplicate data


There are two main methods to deduplicate redundant data: inline and post-processing deduplication. The
backup environment will dictate the method.
Inline deduplication analyzes data as a backup system ingests it. Redundancies are removed as the data is
written to backup storage. Inline dedupe requires less backup storage but can cause bottlenecks. Storage
array vendors recommend that users turn off their inline data deduplication tools for high-performance
primary storage.
Post-processing dedupe is an asynchronous backup process that removes redundant data after it is written
to storage. Duplicate data is removed and replaced with a pointer to the first iteration of the block. The
post-processing approach gives users the flexibility to dedupe specific workloads and quickly recover the
most recent backup without hydration. The tradeoff is a larger backup storage capacity than is required
with inline deduplication.

File-level vs. block-level deduplication


Data deduplication generally operates at the file or block level.
File-level data deduplication compares a file to be backed up or archived with copies that are already
stored. This is done by checking its attributes against an index. If the file is unique, it is stored and the
index is updated; if not, only a pointer to the existing file is stored. The result is only one instance of the
file being saved. Subsequent copies are replaced with a stub that points to the original file.
Block-level deduplication looks within a file and saves unique iterations of each block. All the blocks are
broken into chunks with the same fixed length. Each chunk of data is processed using a hash algorithm,
such as MD5 or SHA-1.
This process generates a unique number for each piece, which is then stored in an index. If a file is
updated, only the changed data is saved, even if only a few bytes of the document or presentation have
changed. The changes don't constitute an entirely new file. This behavior makes block deduplication more
efficient than file deduplication. However, block deduplication takes more processing power and uses a
larger index to track the individual pieces.
Variable-length deduplication is an alternative that breaks a file system into chunks of various sizes,
letting the deduplication effort achieve better data reduction ratios than fixed-length blocks. However, it
also produces more metadata and runs slower.

Benefits and a drawback to deduplication


Hash collisions are a potential problem with deduplication. When a piece of data receives a hash number,
that number is then compared with the index of other existing hash numbers. If that hash number is
already in the index, the piece of data is considered a duplicate and does not need to be stored again.
Otherwise, the new hash number is added to the index and the new data is stored.
In rare cases, the hash algorithm may produce the same hash number for two different chunks of data.
When a hash collision occurs, the system won't store the new data because it sees that its hash number
already exists in the index. This is called a false positive, and it can result in data loss. Some vendors
combine hash algorithms to reduce the possibility of a hash collision. Some vendors are also examining
metadata to identify data and prevent collisions.
Benefits of deduplication include the following:
● a reduced data footprint;
● lower bandwidth consumption when copying data associated with remote backups, replication
and disaster recovery;
● longer retention periods;
● faster recovery time objectives; and
● reduced tape backups.

A cloud backup service offers the use of shared software-defined storage, which is managed like a virtual
resource. This type of virtual architecture enables providers to create a large pool of storage and parcel it
out among many customers.
To use these managed services, there is no need for a dedicated physical or virtual server, off-site
facilities for storing tape backup, or expensive tape drives with dedicated backup software.

Virtualized storage

Software-defined storage enables cloud providers to manage storage at the byte level, and employ multi-
tenant architectures to ensure each account is completely separated from others. This helps isolate data
belonging to different customers.
Customers of cloud backup services can store their frequently used data in several geographic locations.
A vendor-owned data center located close to your office, for example, can provide fast access.

Backup targets and access methods

Cloud storage providers, including Amazon Web Services (AWS) and Microsoft Azure, offer
Infrastructure as a Service (IaaS) models that let you create servers in the cloud as backup targets.
Most cloud backup services provide dedicated solutions that look like network drives to software and
users. While this option provides a high level of flexibility, it also comes with additional fees. You might
also incur additional costs needed to protect the data stored in these services.
Cloud customers can also employ a local storage resource, such as a network-attached storage (NAS)
device, to serve as a middleman for backups. This type of resource can store frequently used files and
then serve them through the faster local network.

Management and storage tiers

Cloud backup vendors usually provide management tools according to customer requirements in terms of
size and demand, security, and changing bandwidth conditions. In some scenarios, variable data retention
requirements are also factored in. This enables cloud providers to automatically drop all files or folders
that are older than the time specified by the administrator.
Storage tiers, like frequently-accessed storage or archival storage, are priced differently. Customers can
define policies and automate how data moves between tiers, to conserve costs for data that is less
frequently used or retained only for compliance purposes.

Assessment questions to the lecture


Bloom’s
Qn No Question Answer Knowledge
Level

1. Duplication of Data is known as b. K1

A. Data Repentance

B. Data Redundancy

C. Data Inconsistency

D. Data base

2. What is data deduplication in the context of data b. K1


backup?

A. The process of encrypting data

B. The process of removing duplicate


copies of data to save storage space
C. The process of creating full backups

D. The process of data recovery

Students have to prepare answers for the following questions at the end of the lecture
Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 List the backup methods. 2 2 K1

2 Differentiate Backup and Restore. 2 2 K1

3 Explain the Backup process and Architecture. 8 2 K2

4 Explain Backup and Restore operations. 8 2 K2

Reference Book

Author(s) Title of the book Page numbers

Web resources
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 4
No 4

Topic Data archive, Uses of replication and its characteristics

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Define Data Archive. K1

LO2 Differentiate Data Archive and Backup. K2

LO3 Explain online and offline data storage. K2

LO4 Discuss the uses of replication and its characteristics. K2

Lecture notes

Data Archive

Data archiving moves data that is no longer actively used to a separate storage device for long-term
retention. Archive data consists of older data that remains important to the organization or must be
retained for future reference or regulatory compliance reasons. Data archives are indexed and have search
capabilities, so files can be located and retrieved.
Archived data is stored on a lower-cost tier of storage, reducing primary storage consumption and the
related costs. An important aspect of a business's data archiving strategy is to inventory its data and
identify what data is a candidate for archiving.
Some archive systems treat archive data as read-only to protect it from modification, while other data
archiving products enable writes as well as reads. For example, WORM -- write once, read many --
technology uses media that is not rewritable.
Data archiving is most suitable for data that must be retained due to operational or regulatory
requirements, such as document files, email messages and possibly old database records.
Data archiving benefits
The greatest benefit of archiving data is it reduces the cost of primary storage. Primary storage is typically
expensive, because a storage array must produce a sufficient level of input/output operations per second
to meet operational requirements for user read/write activity. In contrast, archive storage costs less,
because it is typically based on a low-performance, high-capacity storage medium. Data archives can be
stored on low-cost hard disk drives (HDDs), tape or optical storage that is generally slower than
performance disk or flash drives.
Archive storage also reduces the volume of data that must be backed up. Removing infrequently accessed
data from the backup data set improves backup and restore performance. Typically, organizations perform
data deduplication on data being moved to a lower storage tier, which reduces the overall storage
footprint and lowers secondary storage costs.

Data archiving vs. backup


Data archives are not to be confused with data backups, which are copies of data. Although both are
considered secondary storage and use a lower-performance, higher-capacity storage medium than primary
storage, they serve different purposes. Archives serve a data retention purpose, whereas backups are used
for data protection and disaster recovery.
Data archives can be thought of as a data repository for data that is infrequently accessed but still readily
available. Backups, on the other hand, are part of a data recovery mechanism that can be used
to restore data in the event that it is corrupted or destroyed. Backup data often consists of important
information that must be restored quickly when lost or deleted.
Online vs. offline data storage
Data archives take several different forms. Some systems make use of online data storage, which places
archive data onto disk systems where it is readily accessible. Archives can be file-based or object storage-
based.
Other archival systems use offline data storage in which archive data is written to tape or other removable
media using data archiving software, rather than being kept online. Because tape can be removed, tape-
based archives consume far less power than disk systems. This translates to lower archive storage costs.
Cloud storage is another possible archive target. Amazon Glacier, for example, is designed for data
archiving. This method is inexpensive but requires an ongoing investment. In addition, costs can grow
over time as more data is added to the storage cloud. Cloud providers usually store archived data on tape
or HDDs.

Uses of Replication and its characteristics

Replication is the process of creating an exact copy of data. Creating one or more replicas of the
production data is one of the ways to provide Business Continuity (BC). These replicas can be used for
recovery and restart operations in the event of data loss. The primary purpose of replication is to enable
users to have designated data at the right place, in a state appropriate to the recovery need. The replica
should provide recoverability and restartability.
Recoverability enables restoration of data from the replicas to the production volumes in the event of data
loss or data corruption. It must provide minimal RPO and RTO for resuming business operations on the
production volumes, while restartability must ensure consistency of data on the replica. This enables
restarting business operations using the replicas.

Assessment questions to the lecture


Bloom’s
Qn
Question Answer Knowledge
No
Level

1. What is the primary purpose of data archiving? b. K1


a) To permanently delete old data

b) To organize data for easy retrieval

c) To preserve data that is no longer actively used

d) To encrypt data for security purposes

2. Which of the following is NOT a common use case for data b. K1


replication?

a) Disaster recovery
b) Load balancing
c) Data migration
d) Data compression
Students have to prepare answers for the following questions at the end of the lecture

Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 Define Data Archive. 2 4 K1

2 Differentiate Data Archive and Backup. 2 4 K2

3 Explain online and offline data storage. 8 4 K2

4 Discuss the uses of replication and its characteristics. 8 4 K2

Reference Book

Author(s) Title of the book Page numbers

G Somasundaram EMC Corporation, Information Storage and 284-285


Management, Wiley, India
Alok Shrivastava
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 5
No 4

Topic Compute based, storage-based, and network-based replication

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Define Storage system based Replication. K1

LO2 Define Network based replication. K2

LO3 Discuss the types of Storage system based replication. K2

Lecture notes

Storage System based Replication


In the storage system based replication, the storage system operating environment performs the
replication process. Unlike Server based replication, server resources such as the CPU and memory, are
not used in the replication process. Consequently, the server running multiple VMs is not burdened by the
replication operations. Storage system-based replication supports both local and remote replication.

Storage System based Local Replication Techniques


● Full Volume Replication (Cloning)
● Pointer based Virtual Replication (Snapshot)
Storage System based Remote Replication Techniques
● Synchronous Replication
● Asynchronous Replication
● Multi-site Replication
In storage system based local replication, the replication is performed within the storage system. In other
words, the source and the target LUNs reside on the same storage system. Local replication enables one to
perform operational recovery in the event of data loss and also provides the support for other business
operations such as backup. The storage system-based local replication can be implemented as full volume
replication (clone), and pointer-based virtual replication (snapshot).

In storage system-based remote replication, the replication is performed between storage systems.
Typically one of the storage systems is in source site and the other system is in remote site for DR
purpose. Data can be transmitted from the source storage system to the target system over a shared or a
dedicated network. Replication between storage systems may be performed in synchronous or
asynchronous modes.

Storage System based Local Replication Techniques


Full Volume Replication (Cloning)
Full volume replication provides the ability to create fully populated point-in-time copies of LUNs within
a storage system. When the replication session is started, an initial synchronization is performed between
the source LUN and the replica (clone). Synchronization is the process of copying data from the source
LUN to the clone. During synchronization process, the replica is not available for any server access. Once
the synchronization is completed, the replica is exactly same as source LUN. The replica can be detached
from the source LUN and it can be made available to another server for business operations. Subsequent
synchronizations involve only a copy of any data that has changed on the source LUN since the previous
synchronization.

Typically after detachment, changes made to both the source and replica can be tracked at some
predefined granularity. This enables incremental resynchronization (source to target) or incremental
restore (target to source). The clone must be the same size as the source LUN.

Pointer based Virtual Replication (Snapshot)


Pointer-based virtual replication (also referred as storage system-based snapshot) is a space optimal
solution when compared to full volume replica. At the time of replication session activation, the target
(snapshot) contains pointers to the location of the data on the source. The snapshot does not contain data
at any time. Therefore, the snapshot is known as a virtual replica. Snapshot is immediately accessible
after the replication session activation. This replication method either uses Copy on First Write (CoFW)
or Redirect on Write (RoW) mechanism. Multiple snapshots can be created from the same source LUN
for various business requirements.

Data on the target is a combined view of unchanged data on the source and data on the save location. The
unavailability of the source device invalidates the data on the target. The target contains only pointers to
the data, and therefore, the physical capacity required for the target is a fraction of the source device. The
capacity required for the save location depends on the amount of the expected data change.

Some pointer-based virtual replication implementation uses redirect on write technology (RoW). RoW
redirects new writes destined for the source LUN to a reserved LUN in the storage pool. Such
implementation is different from CoFW, where the writes to the source LUN are held until the original
data is copied to the save location to preserve the point-in-time replica. There is always a need to perform
a lookup to determine whether data is on the source LUN or save location, which causes snapshot reads to
be slower than source LUN reads. In the case of a RoW snapshot, the original data remains where it is,
and is therefore read from the original location on the source LUN.
Storage System based Remote Replication Techniques
Synchronous Remote replication
Storage-based remote replication solution can avoid downtime by enabling business operations at remote
sites. Storage-based synchronous remote replication provides near zero RPO where the target is identical
to the source at all times. In synchronous replication, writes must be committed to the source and the
remote target prior to acknowledging “write complete” to the production server. Additional writes on the
source cannot occur until each preceding write has been completed and acknowledged. This ensures that
data is identical on the source and the target at all times. Further, writes are transmitted to the remote site
exactly in the order in which they are received at the source. Therefore, write ordering is maintained and
it ensures transactional consistency when the applications are restarted at the remote location. Most of the
storage systems support consistency group, which allows all LUNs belonging to a given application,
usually a database, to be treated as a single entity and managed as a whole. This helps to ensure that the
remote images are consistent. As a result, the remote images are always restartable copies.

Asynchronous Remote Replication


It is important for an organization to replicate data across geographical locations in order to mitigate the
risk involved during disaster. If the data is replicated (synchronously) between sites and the disaster
strikes, then there would be a chance that both the sites may be impacted. This leads to data loss and
service outage. Replicating data across sites which are 1000s of kilometers apart would help organization
to face any disaster. If a disaster strikes at one of the regions then the data would still be available in
another region and the service could move to the location. Asynchronous replication enables to replicate
data across sites which are 1000s of kilometers apart.

In asynchronous remote replication, a write from a production server is committed to the source and
immediately acknowledged to the server. Asynchronous replication also mitigates the impact to the
application’s response time because the writes are acknowledged immediately to the server. This enables
to replicate data over distances of up to several thousand kilometers between the source site and the
secondary site (remote locations). In asynchronous replication, the server writes are collected into buffer
(delta set) at the source. This delta set is transferred to the remote site in regular intervals. Therefore,
adequate buffer capacity should be provisioned to perform asynchronous replication. In asynchronous
replication, RPO depends on the size of the buffer, the available network bandwidth, and the write
workload to the source. This replication can take advantage of locality of reference (repeated writes to the
same location). If the same location is written multiple times in the buffer prior to transmission to the
remote site, only the final version of the data is transmitted. This feature conserves link bandwidth.

Multi-site Remote Replication


In a two-site synchronous replication, the source and target sites are usually within a short distance.
Therefore, if a regional disaster occurs, both the source and the target sites might become unavailable.
This can lead to extended RPO and RTO because the last known good copy of data would need to come
from another source, such as an offsite tape. A regional disaster will not affect the target site in a two-site
asynchronous replication because the sites are typically several hundred or several thousand kilometers
apart. If the source site fails, production can be shifted to the target site, but there is no further remote
protection of data until the failure is resolved.

Introduction to Network Based Replication


In the Network Based replication, the replication occurs at the network layer between the server and the
storage systems. By offloading replication from server and storage systems, network-based replication
can work across a large number of server platforms and storage systems, making it ideal for highly
heterogeneous environments. One of the most widely used Network based replication technique is
the Continuous Data Protection (CDP).

Assessment questions to the lecture

Bloom’s
Qn
Question Answer Knowledge
No
Level

1. is the process of creating replicas of information a. K1


assets at remote sites

a.Remote replication

b.Local replication

c.Both

d.None
2. Remote replicas help organizations mitigate the risks True K1
associated with regionally driven outages resulting from
natural or human-made disasters.

True

False

Students have to prepare answers for the following questions at the end of the lecture

Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 Define Storage system based Replication. 2 2 K1

2 Define Network based Replication. 2 2 K1

3 Discuss the types of Storage system based replication. 8 2 K2

Reference Book

Author(s) Title of the book Page numbers

G Somasundaram EMC Corporation, Information Storage and 286-290


Management, Wiley, India
Alok Shrivastava
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 7
No 4

Topic Data migration-Types of Migration-Data migration process.

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Describe Data migration. K1

LO2 Discuss the types of Data migration. K2

LO3 Explain the Data migration process. K2

Data migration

Data migration is the transfer of the existing historical data to new storage, system, or file format. This
process is not as simple as it may sound. It involves a lot of preparation and post-migration activities
including planning, creating backups, quality testing, and validation of results. The migration ends only
when the old system, database, or environment is shut down.
Usually, data migration comes as a part of a larger project such as
● legacy software modernization or replacement,
● the expansion of system and storage capacities,
● the introduction of an additional system working alongside the existing application,
● the shift to a centralized database to eliminate data silos and achieve interoperability,
● moving IT infrastructure to the cloud, or
● merger and acquisition (M&A) activities when IT landscapes must be consolidated into a single
system.

Types of data migration

There are six commonly used types of data migration. However, this division is not strict. A particular
case of the data transfer may belong, for example, to both database and cloud migration or involve
application and database migration at the same time.
Storage migration
Storage migration occurs when a business acquires modern technologies discarding out-of-date
equipment. This entails the transportation of data from one physical medium to another or from a physical
to a virtual environment. Examples of such migrations are when you move data
● from paper to digital documents,
● from hard disk drives (HDDs) to faster and more durable solid-state drives (SSDs), or
● from mainframe computers to cloud storage.

Database migration
A database is not just a place to store data. It provides a structure to organize information in a specific
way and is typically controlled via a database management system (DBMS).

So, most of the time, database migration means


● an upgrade to the latest version of DBMS (so-called homogeneous migration),
● a switch to a new DBMS from a different provider — for example, from MySQL to PostgreSQL
or from Oracle to MSSQL (so-called heterogeneous migration)

Application migration

When a company changes an enterprise software vendor — for instance, a hotel implements a
new property management system or a hospital replaces its legacy EHR system — this requires moving
data from one computing environment to another. The key challenge here is that old and new
infrastructures may have unique data models and work with different data formats.

Data center migration


A data center is a physical infrastructure used by organizations to keep their critical applications and data.
Put more precisely, it’s the very dark room with servers, networks, switches, and other IT equipment. So,
data center migration can mean different things: from relocation of existing computers and wires to other
premises to moving all digital assets, including data and business applications to new servers and
storages.

Business process migration


This type of migration is driven by mergers and acquisitions, business optimization, or reorganization to
address competitive challenges or enter new markets. All these changes may require the transfer of
business applications and databases with data on customers, products, and operations to the new
environment.

Cloud migration
Cloud migration is a popular term that embraces all the above-mentioned cases, if they involve moving
data from on-premises to the cloud or between different cloud environments.

Data migration process


The data migration project goes through the same key phases — namely
● planning,
● data auditing and profiling,
● data backup,
● migration design,
● execution,
● testing, and
● post-migration audit.

Planning: create a data migration plan and stick to it


Data migration is a complex process, and it starts with the evaluation of existing data assets and careful
designing of a migration plan. The planning stage can be divided into four steps.

Step 1 Refine the scope. The key goal of this step is to filter out any excess data and to define the
smallest amount of information required to run the system effectively. So, you need to perform a high-
level analysis of source and target systems, in consultation with data users who will be directly impacted
by the upcoming changes.

Step 2 Assess source and target systems. A migration plan should include a thorough assessment of the
current system’s operational requirements and how they can be adapted to the new environment.

Step 3 Set data standards. This will allow your team to spot problem areas across each phase of the
migration process and avoid unexpected issues at the post-migration stage.

Step 4 Estimate budget and set realistic timelines. After the scope is refined and systems are evaluated,
it’s easier to select the approach (big bang or trickle), estimate resources needed for the project, set
schedules, and deadlines. According to Oracle estimations, an enterprise-scale data migration project lasts
six months to two years on average.

Assessment questions to the lecture

Bloom’s
Qn
Question Answer Knowledge
No
Level

1. What is data migration? b. K1

a) Deleting old data from a system


b) Transferring data from one system or storage
environment to another
c) Backing up data for future use
d) Encrypting data for security purposes

2. What is the first step in the data migration process? c. K1

a) Testing and Validation


b) Data Profiling and Cleansing
c) Planning and Assessment
d) Execution and Monitoring

3. Which factor is crucial for the success of a data migration c. K1


project?

a) Skipping data validation


b) Ignoring stakeholder feedback
c) Establishing clear objectives and requirements
d) Using outdated migration tools

Students have to prepare answers for the following questions at the end of the lecture
Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 Describe Data migration. 2 2 K1


2 Discuss the types of Data migration. 2 2 K1

3 Explain the Data migration process. 8 2 K2

Reference Book

Author(s) Title of the book Page numbers

Web resources
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 8
No 4

Topic Disaster recovery as a service

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Define DRaaS. K1

LO2 Differentiate DRaaS and BaaS. K2

LO3 Explain DRaaS. K2

Lecture notes

Disaster recovery as a service (DRaaS) is a cloud computing service model that allows an organization
to back up its data and IT infrastructure in a third party cloud computing environment and provide all the
DR orchestration, all through a SaaS solution, to regain access and functionality to IT infrastructure after
a disaster. The as-a-service model means that the organization itself doesn’t have to own all the resources
or handle all the management for disaster recovery, instead relying on the service provider.

Disaster recovery planning is critical to business continuity. Many disasters that have the potential to
wreak havoc on an IT organization have become more frequent in recent years:
● Natural disasters such as hurricanes, floods, wildfires and earthquakes
● Equipment failures and power outages
● Cyberattacks

DRaaS works by replicating and hosting servers in a third-party vendor’s facilities versus in the physical
location of the organization that owns the workload. The disaster recovery plan is executed on the third-
party vendor’s facilities in the event of a disaster that shuts down a customer’s site. Organizations may
purchase DRaaS plans through a traditional subscription model or a pay-per-use model that allows them
to pay only when disaster strikes. As-a-service solutions vary in scope and cost—organizations should
evaluate potential DRaaS providers according to their own unique needs and budget.

DRaaS can save organizations money by eliminating the need for provisioning and maintaining an
organization’s own off-site disaster recovery environment. However, organizations should evaluate and
understand service level agreements. For instance, what happens to recovery times if both the provider
and customer are affected by the same natural disaster, such as a large hurricane or earthquake. Different
DRaaS providers have different policies on prioritizing which customers get help first in a large regional
disaster or allowing customers to perform their own disaster recovery testing.

DRaaS vs. backup as a service (BaaS)

With disaster recovery as a service, the service provider moves an organization’s computer processing to
its cloud infrastructure in the event of a disaster. This way, the business can continue to operate, even if
the original IT infrastructure is totally destroyed or held hostage. This differs from backup as a service,
where only the data, but not the ability to process the data, is duplicated by a third-party provider.
Because BaaS is only protecting the data, and not the infrastructure, it is typically less expensive than
DRaaS. BaaS can be a good solution for companies that need to archive data or records for legal reasons,
but most organizations who use BaaS will want to combine it with another disaster recovery tool to
ensure business continuity.

Planning for disaster and getting the help you need is something every business needs to consider.
Whatever option you choose, a disaster recovery plan is essential for business continuity, and
organizations are increasingly turning to DRaaS.

Assessment questions to the lecture

Bloom’s
Qn
Question Answer Knowledge
No
Level

1. What is Disaster Recovery as a Service (DRaaS)? b. K1

a) A physical backup solution for data centers


b) A cloud-based disaster recovery solution
c) A manual process for restoring data after a disaster
d) A software tool for data encryption

2. Which of the following is a key benefit of Disaster c. K1


Recovery as a Service (DRaaS)?

a) High upfront costs


b) Complex implementation
c) Reduced recovery time objectives (RTO)
d) Limited scalability

3. What role does data encryption play in Disaster Recovery c. K1


as a Service (DRaaS)?

a) It increases recovery time objectives (RTO)


b) It reduces data redundancy
c) It ensures data security during replication and storage
d) It decreases reliance on cloud resources
Students have to prepare answers for the following questions at the end of the lecture

Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 Define DRaaS. 2 2 K1

2 Differentiate DRaaS and BaaS. 2 2 K1

3 Explain DRaaS in detail. 8 2 K2

Reference Book

Author(s) Title of the book Page numbers

G Somasundaram EMC Corporation, Information Storage and 305-310


Management, Wiley, India
Alok Shrivastava
Unit
BACKUP, ARCHIVE AND REPLICATION Lecture No 9
No 4

Topic Cloud-based and mobile device backup

Bloom’s Knowledge
Learning Outcome (LO) At the end of this lecture, students will be able to
Level

LO1 Define Cloud based backup. K1

LO2 Discuss Cloud backup usecases and approaches. K2

LO3 Explain mobile device backup. K2

Lecture notes

Cloud backup, also known as online backup or remote backup, is a strategy for sending a copy of a
physical or virtual file or database to a secondary, off-site location for preservation in case of equipment
failure, site catastrophe or human malfeasance. The backup server and data storage systems are usually
hosted by a third-party cloud or SaaS provider that charges the backup customer a recurring fee based on
storage space or capacity used, data transmission bandwidth, number of users, number of servers or
number of times data is retrieved.
Implementing cloud data backup can help bolster an organization's data protection, business continuance
and regulatory compliance strategies without increasing the workload of IT staff. The labor-saving benefit
can be significant and enough of a consideration to offset some of the additional costs associated with
cloud backup, such as data transmission charges.
Most cloud subscriptions run on a monthly or yearly basis. Initially used mainly by consumers and home
offices, online backup services are now commonly used by SMBs and larger enterprises to back up some
forms of data. For larger companies, cloud data backup can serve as a supplementary form of backup.
What is the cloud?
Cloud computing is a general term that refers to hosted resources and services that are delivered over the
internet. Different from traditional web hosting, the services on the cloud are sold on demand, offered in
an elastic manner -- meaning the customer can use as much or as little of the service as needed -- and
managed completely by the service provider. Additionally, a cloud can be private or public. A public
cloud sells services to anyone on the internet, such as how AWS operates, while a private cloud supplies
hosted services to a limited number of users within the business.

Cloud backup use cases and approaches


In an organization's data center, a backup application copies data and stores it on different media or
another storage system for easy access in the event of a recovery situation Although there are multiple
options and approaches to off-site backup, cloud backup serves as the off-site facility for many
organizations. In an enterprise, the company might own the off-site server if it hosts its own private cloud
service, but the chargeback method would be similar if the company uses a service provider to manage
the cloud backup environment and receives a regular bill for backup storage and services.
There are a variety of approaches to cloud backup, with available services that can easily fit into an
organization's existing data protection process. Varieties of cloud backup include the following:
 Backing up directly to the public cloud. One way to store organizational workloads is by
duplicating resources in the public cloud. This method entails writing data directly to cloud
providers, such as AWS, Google Cloud or Microsoft Azure. The organization uses its own
backup software to create the data copy to send to the cloud storage service. The cloud storage
service then provides the destination and safekeeping for the data, but it doesn't specifically
provide a backup application. In this scenario, it's important that the backup software is capable
of interfacing with the cloud's storage service. Additionally, with public cloud options, IT
professionals might need to look into supplemental data protection procedures, such as data
encryption as well as identity and access management to secure backed up data.
 Backing up to a service provider. In this scenario, an organization writes data to a cloud service
or SaaS provider that offers backup services in a managed data center. The backup software that
the company uses to send its data to the service might be provided as part of the service, or the
service might support specific commercially available backup applications.
 Choosing a cloud-to-cloud (C2C) backup. These services are among the newest offerings in the
cloud backup arena. They specialize in backing up data that already lives in the cloud, either as
data created using a SaaS application or as data stored in a cloud backup service. As its name
suggests, a C2C backup service copies data from one cloud to another cloud. The cloud-to-cloud
backup service typically hosts the software that handles this process.
 Using online cloud backup systems. There are also hardware alternatives that facilitate backing
up data to a cloud backup service. These appliances are all-in-one backup machines that include
backup software and disk capacity, along with the backup server. The appliances are about as
close to plug-and-play as backup gets, and most of them also provide a seamless link to one or
more cloud backup services or cloud providers. The list of vendors that offer backup appliances
that include cloud interfaces is long, with Quantum, Unitrends, Arcserve, Rubrik, Cohesity, Dell
EMC, StorageCraft and Asigra active in this arena. These appliances typically retain the most
recent backup locally, in addition to copying it to the cloud backup provider, so any required
recoveries can be made from the local backup copy, saving time and transmission costs.

Although the steps can vary based on backup method or type, this is the basic process for cloud backups.

When an organization engages a cloud backup service, the first step is to complete a full backup of the
data that must be protected. This initial backup can sometimes take days to finish uploading over a
network as a result of the large volume of data being transferred. In a 3-2-1 backup strategy, where an
organization has three copies of data on two different media, at least one copy of the backed up data
should be sent to an off-site backup facility so that it's accessible even if on-site systems are unavailable.
Using a technique called cloud seeding, a cloud backup vendor sends a storage device -- such as a hard
drive or tape cartridge -- to its new customer, which then backs up the data locally onto the device and
returns it to the provider. This process removes the need to send the initial data over the network to the
backup provider. One example of a device that employs this technique is AWS Snowball Edge.
If the amount of data in the initial backup is substantial, the cloud backup service might provide a full
storage array for the seeding process. These arrays are typically small network-attached storage (NAS)
devices that can be shipped back and forth relatively easily. After the initial seeding, only changed data is
backed up over the network.
How data is restored
Cloud backup services are typically built around a client software application that runs on a schedule
determined by the purchased level of service and the customer's requirements. For example, if the
customer has contracted for daily backups, the application collects, compresses, encrypts and transfers
data to the cloud service provider's servers every 24 hours. To reduce the amount of bandwidth consumed
and the time it takes to transfer files, the service provider might only provide incremental backups after
the initial full backup.
Cloud backup services often include the software and hardware necessary to protect an organization's
data, including applications for Microsoft Exchange and SQL Server. Whether a customer uses its own
backup application or the software the cloud backup service provides, the organization uses that same
application to restore backed up data. Restorations could be on a file-by-file basis, by volume or a full
restoration of the complete backup. More granular file-by-file restoration is typically the preferred method
because it enables a business to quickly recover individual lost or damaged files rather than take the time
and risk in restoring entire volumes.
If the volume of data to be restored is very large, the cloud backup service might ship the data on a
complete storage array that the customer can hook up to its servers to recover its data. This is, in effect, a
reverse seeding process. Restoring a large amount of data over a network can take an unacceptably long
time, depending on the organization's recovery time objective.
A key feature of cloud backup restorations is that they can be done anywhere, from nearly any kind of
computer. For example, an organization could recover its data directly to a disaster recovery site in a
different location if its primary data center is unavailable.
Mobile Device Backup:

Mobile device backup involves creating copies of data stored on smartphones, tablets, or other mobile
devices to prevent data loss in case of device damage, loss, or theft. Key aspects of mobile device backup
include:

 Data Types: Mobile device backup typically includes contacts, photos, videos, messages, app
data, and other user-generated content.

 Backup Methods: Mobile device backup can be performed using various methods, including
built-in backup features provided by the device's operating system, third-party backup apps, or
cloud-based backup services specifically designed for mobile devices.

 Automatic Backup: Many mobile devices offer automatic backup options that regularly back up
data to the cloud or other storage locations, ensuring that the backup is up-to-date.

 Restore Options: In addition to backup, mobile devices often provide easy-to-use restore options
to recover data onto a new or reset device.

 Data Encryption: To ensure data security, mobile device backup solutions may incorporate
encryption techniques to protect data during transmission and storage.

Examples of mobile device backup solutions include iCloud for iOS devices, Google Backup for Android
devices, and various third-party backup apps available on app stores.

Assessment questions to the lecture

Bloom’s
Qn
Question Answer Knowledge
No
Level

1. What is cloud-based backup? b. K1

a) Storing data on physical backup tapes


b) Backing up data to a remote server over the internet
c) Copying data to an external hard drive
d) Replicating data on-premises

2. What is mobile device backup? d. K1

a) Storing data on physical backup tapes


b) Backing up data to a computer's hard drive
c) Copying data to an external storage device
d) Saving data from mobile devices to a remote server or
cloud

3. Which of the following data types are commonly backed up c. K1


from mobile devices?

a) System files only


b) Applications only
c) Contacts, photos, videos, and app data
d) Operating system files only

Students have to prepare answers for the following questions at the end of the lecture

Marks CO Bloom’s
Qn
Question Knowledge
No
Level

1 Define Cloud based backup. 2 2 K1

2 Discuss Cloud backup usecases and approaches. 8 2 K1

3 Explain mobile device backup. 8 2 K2

Reference Book

Author(s) Title of the book Page numbers

Web resources

You might also like