Full Text 01
Full Text 01
ElectricalEngineering
092014
Revanth Vemulapalli
Ravi Kumar Mada
Contact Information:
Author(s):
Revanth Vemulapalli,
Ravi Kumar Mada.
E-mail:
[email protected],
[email protected].
University advisor(s):
Dr. Dragos Ilie,
Department of Communication Systems.
University examiner(s):
Prof. Kurt Tutschku,
Department of Communication Systems.
School of Computing
Blekinge Institute of Technology Internet : www.bth.se/com
SE-371 79 Karlskrona Phone : +46 455 38 50 00
Sweden Fax : +46 455 38 50 57
Abstract
I would like to express gratitude to my supervisor Dr. Dragos Ilie who intro-
duced the concept of virtualization to us and for his encouragement and support
through valuable suggestions when required. Furthermore, I would like to thank
Prof. Dr. Kurt Tutschku for his motivation and useful suggestions through the
learning process of this master thesis. I would like to thank Dr. Patrik Arlos for
lending us hardware for our experiments.
I am fortunate to have loving and caring parents who supported my stay and
education against all odds. Without their support I couldn‘t have studied in
Sweden. In my daily work I have been blessed with a friendly and cheerful group
of fellow students (Chaitu, Gautam, Venu, Tushal, etc). I will forever be grateful
for all your love and help.
–Revanth Vemulapalli
iii
iv
Acknowledgments
I would like to thank Dr. Patrik Arlos for providing required equipment and
environment. I appreciate the cooperation of my partner Revanth.V for being
consistent towards this research work.
I would like to thank my friends who have been great support. Last but not
the least; I would also like to thank my family for their endless support.
–Ravi Kumar M
v
vi
Contents
Abstract i
Acknowledgments iii
Acknowledgments v
List of Figures xi
List of Tables xv
1 INTRODUCTION 1
1.1 Advantages of virtualization . . . . . . . . . . . . . . . . . . . . . 2
1.2 Components of Virtualization . . . . . . . . . . . . . . . . . . . . 2
1.3 Types of Virtualization . . . . . . . . . . . . . . . . . . . . . . . . 3
1.3.1 Full Virtualization . . . . . . . . . . . . . . . . . . . . . . 3
1.3.2 Para Virtualization . . . . . . . . . . . . . . . . . . . . . . 3
1.3.3 Hardware Assisted Virtualization . . . . . . . . . . . . . . 4
1.4 Virtual Machine Migration . . . . . . . . . . . . . . . . . . . . . . 4
1.5 Problem Statement . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.6 Advantages of Live Migration over WAN . . . . . . . . . . . . . . 7
vii
3.2.3 Multiple Replication Transports . . . . . . . . . . . . . . . 23
3.2.4 Three Way Replication . . . . . . . . . . . . . . . . . . . . 24
3.2.5 Efficient Synchronization . . . . . . . . . . . . . . . . . . . 24
3.2.6 XEN with DRBD . . . . . . . . . . . . . . . . . . . . . . . 25
3.3 iSCSI Block Storage . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.3.1 iSCSI Architecture . . . . . . . . . . . . . . . . . . . . . . 27
3.4 Network File Storage (NFS) . . . . . . . . . . . . . . . . . . . . . 29
5 Experiment Results 40
5.1 Input Output per Second Performance . . . . . . . . . . . . . . . 41
5.1.1 No delay (0.02 ms) . . . . . . . . . . . . . . . . . . . . . . 41
5.1.2 Inter-Europe (30ms Delay) . . . . . . . . . . . . . . . . . . 42
5.1.3 Trans-Atlantic (90ms Delay) . . . . . . . . . . . . . . . . . 42
5.1.4 Trans-Pacific (160ms Delay) . . . . . . . . . . . . . . . . . 43
5.2 I/O Throughput during Migration . . . . . . . . . . . . . . . . . . 44
5.2.1 Campus Network (0.02 ms) . . . . . . . . . . . . . . . . . 45
5.2.2 Inter-Europe (30ms Delay) . . . . . . . . . . . . . . . . . . 45
5.2.3 Trans-Atlantic (90ms Delay) . . . . . . . . . . . . . . . . . 46
5.2.4 Trans-Pacific (160ms Delay) . . . . . . . . . . . . . . . . . 46
5.3 Maximum Response Time (MRT) . . . . . . . . . . . . . . . . . . 48
5.3.1 No delay (0.02 ms) . . . . . . . . . . . . . . . . . . . . . . 48
viii
5.3.2 Inter-Europe (30ms Delay) . . . . . . . . . . . . . . . . . . 48
5.3.3 Trans-Atlantic (90ms Delay) . . . . . . . . . . . . . . . . . 49
5.3.4 Trans-Pacific (160ms Delay) . . . . . . . . . . . . . . . . . 50
5.4 Network Load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
Appendix 56
ix
x
List of Figures
1.1 Virtualization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Type-1 and Type-2 Hypervisor . . . . . . . . . . . . . . . . . . . 3
1.3 Types of Virtualization . . . . . . . . . . . . . . . . . . . . . . . . 4
1.4 Xen Live Migration [9] . . . . . . . . . . . . . . . . . . . . . . . . 5
xi
xii
List of Equations
xiii
xiv
List of Tables
xv
xvi
Acronyms
ABSS Activity Bases Sector Synchronization MBPS Mega Bytes per Second
AVG average page dirty rate MIPv6 Mobile Internet Protocol version 6
CBR Content Based Redundancy MRT Maximum Response Time
Challenge-Handshake Authentication
CHAP Protocol NAS Network Attached Storage
CPU Central Processing Unit NFS Network File Storage
DBT Dirty Block Tracking OS Operating System
DDNS Dynamic Domain Name Service PBA Proxy Binding Acknowledgement
DNS Domain Name Service PBU Proxy Binding Unit
Proxy Mobile Internet Protocol
DRBD Distributed Replicated Block Device PMIPv6 version 6
HIST history based page dirty rate RDMA Remote Direct Memory Access
IM Incremental Migration RPC Remote Procedure Call
IO Input Output RPD Rapid Page Dirtying
IOPS Input Outputs per Second SAN Storage Area Network
IPSec Internet Protocol Security SFHA Super-Fast Hash Algorithm
IPv4 Internet Protocol version 4 TCP Transmission Control Protocol
IPv6 Internet Protocol version 6 TOE TCP offload engine
internet Small Computer System
iSCSI Interface VM Virtual Machine
KVM Kernel-based Virtual Machine VMM Virtual Machine Monitor
LAN Local Area Network VMTC Virtual Machine Traffic Controller
LMA Local Mobility Anchor VPN Virtual Private Network
MAG Mobile Access Gateway WAN Wide Area Network
xviii
Chapter 1
INTRODUCTION
This report is divided into six chapters. The concepts of virtualization and
1
Chapter 1. INTRODUCTION 2
advantages of VM migration over WAN are discussed in chapter one. Chapter two
describes our aim, research question and related work. We introduce the XEN
hypervisor, the disk replication scheme DRBD and shared storage approaches like
iSCSi and NFS in chapter three. We describe our research methodology and ex-
periment setups with detail description of hardware used in our experiment along
with delay characteristics of WAN in chapter four. The analysis of the results
observed during the experiment are presented in chapter five. We concluded our
thesis and discussed possible future work in chapter 6.
and every virtual machine to work. Virtual box, VMWare player and VMware
workstation are examples of type-2 hypervisors.
The guest is the virtual host that runs above the virtualization layer. The
virtual host has its own operating system (OS) and applications. This guest
operating system can be migrated from one physical host to other physical host.
migration, the virtual machine is suspended at the source node before migrating
it to the destination node where it is resumed. Most of the OS state can be
preserved during this migration.
In live migration [9] the hosted operating system is migrated along with CPU,
memory and disk state from source to destination while the hosted OS is still
in running state without losing the active network connectivity. The disk state
migration is not necessary in the case of using shared storages like Network At-
tached Storage (NAS) or Storage Area Network (SAN). Among three migration
techniques, live migration is best suited to reduce notable downtime of services
running on the virtual machine [9]. By using live migration, load on physical
machines hosting several virtual machines can be decreased to a large extent [22].
Live migration may increase the total migration time of the virtual machines
running server applications.
In order to perform live migration, hypervisor needs to move the memory
state of the virtual machine from the source to the destination physical machine.
The crucial method used for migrating memory state is known as pre-copy, and
Chapter 1. INTRODUCTION 6
is clearly explained in [8] and [9]. In pre-copy phase the memory state of vir-
tual machine is copied to destination host in iterations. Unmodified or unused
pages are moved in first round and modified pages are moved in next nth round.
Hypervisor maintains dirty bitmap to track modified pages. If pre-determined
bandwidth is reached or when bandwidth range is below 256kb, pre-copy phase
is terminated [4] and stop and copy phase is initiated. In stop and copy phase
the virtual machine is paused on source host and modified pages are copied to
destination host. The virtual host is resumed at the destination host and starts
working [3] as usual. Xen live migration time line is shown in figure 1.4 [9].
Various disk state migration and network state migration schemes proposed by
various researchers are discussed in related work section of chapter 2.
commonly to two or more hosts. If a virtual machine use private storage it requires
the virtual disk to be migrated along with memory, CPU and network states. This
increases the downtime and total migration time of the virtual machine. Frequent
disk migrations also leads to high network overhead due to their size. Therefore
shared storage is utilized to eliminate the need of disk state migration when a
virtual machine is migrated. This will reduce the total migration and downtimes.
However, when migrated over WAN, the performance of virtual machine may still
be effected due to the high latency in the network. This is because the virtual
machine uses network to access disk which is located at another geographical
location. To our best knowledge the disk I/O performance of virtual machine
while migrating over WAN was not fully explored earlier. Our thesis work may
look similar to [2], the difference between both the works is discussed in section
2.3.
The primary purpose of our thesis is to investigate disk performing, during the
live migration of a virtual machine over the WAN using different shared storage
techniques. We will analyse the performance of iSCSI, NFS and DRBD storage
techniques and recommend the best technique for specific scenarios.
In this chapter we briefly described our aim and objectives along with research
questions. We discussed the related work conducted by various researchers on
network and disk state migrations.
9
Chapter 2. Aim, Objectives, Research Questions & Related Work 10
RQ1:
What are the strategies proposed by various researchers to support live
virtual machine migration over WAN/Internet?
RQ2:
What are the various disc migration schemes proposed by various researchers?
RQ3:
How efficient is the Distributed Replicated Block Device (DRBD) to mirror
disc data in different locations?
RQ4:
What is the I/O performance when migrating virtual machines, over a WAN
using DRBD?
RQ5:
What is the I/O performance when migrating virtual machines with shared
storage solutions over a WAN?
RQ6:
Among distributed and shared storages which solution performs better while
migrating over WAN?
memory pages are migrated to destination along with CPU state. The third
stage is the combination of push and pull phase. Here the virtual machine is
resumed and according to the details in dirty bitmap, the source push the dirty
blocks and the destination machine pulls them. This process reduce the total
migration time of the virtual machine.
grating disk state over WAN, which created two copies of data disks by mirroring.
DRBD supports asynchronous, semi synchronous and full synchronous replication
modes. All the three replication modes are discussed in chapter 3.2.2. They ex-
pected that using DRBD in asynchronous mode gives better performance. They
constructed a laboratory test bed with 50ms RTT delay to emulate WAN. They
compared the performance of DRBD with high latency Internet Small Computer
System Interface (iSCSI) and concluded DRBD test is 2.5 times faster than re-
mote iSCSI test. The downtime observed during the virtual machine’s migration
is only 1.9 seconds. While coming to read statistics, DRBD’s performance is
better when compared to remote iSCSI. But there is no significant difference be-
tween the performance of DRBD and local iSCSI. However, the authors admit
that there are some inconsistencies in their results, which they cannot account
for. Our study may look similar to this work. In our thesis we include NFS
as a shared storage solution and we also measure the performance for different
latencies. In this authors analyzed the virtual machines disk performance based
Chapter 2. Aim, Objectives, Research Questions & Related Work 13
on migration time, http mean request time and I/O read statistics. They used
DRBD asynchronous replication algorithm to replicate data between nodes. We
analyzed the disk I/O performance using different performance metrics discussed
in section 4.6 by conducting both read and write I/O tests. We used DRBD
synchronous replication model which is more secure than asynchronous protocol
used in [2]. Our experiment results confirm that DRBD outperformed iSCSI and
NFS while the virtual machine is migrated over WAN.
Wood et al. [11] discussed solutions for the limitations faced by virtual
machines migrating over WAN. They experimented with an architecture called
CloudNet that is distributed across datacenters in the United States separated
by 1200km. Their architecture reduced network bandwidth utilization by 50%,
and memory migration and pause time by 30 to 70%. The authors used the
DRBD disk replication system to replicate disk in both source and destination.
Firstly DRBD brings the remote disk to consistent stage using synchronization
demon. When both the disks are synchronized and stable, DRBD’s synchronous
replication protocol is switched and the modified data is placed in TCP buffer
to transmit towards destination disk. When the write state is confirmed on des-
tination the synchronous mode is completed. The memory and CPU modes are
migrated after disk migration stage. Content Based Redundancy (CBR) [11]
block technique is used to save bandwidth. CBR splits disk blocks and memory
pages into fixed size finite blocks and uses Super-Fast Hash Algorithm (SFHA)
to generate hash bases on their content. This hash is used to compare previously
sent blocks. This solution saved 20 GB bandwidth when compared to the pro-
cess of migrating full disk. It also reduced memory transfer time by 65%. They
designed a scheme to retain network state using layer-2 VPN solution. In this
experiment authors analyzed the network bandwidth utilized for virtual machines
migration along with memory migration and pause times. On the other hand, in
our experiment, we focused on the disk I/O performance of DRBD, iSCSI and
NFS when the VM is migrating over WAN along with network utilization. From
our experiments we can say that iSCSI and NFS consumed more than 75% of
network bandwidth than DRBD.
Akoush et al. discussed the parameters [19] which effect the live migration
of virtual machines. Total migration time and downtime are the two important
performance parameters used to analyze virtual machines performance during
migration. Total migration time is the time required to move virtual machines
from one physical host to another physical host. Downtime is a period of time
where virtual machine doesn’t run. Page dirtying rate and network bandwidth
are the factors which affect the total migration time and downtime. The authors
implemented two simulation models, namely average page dirty rate (AVG) and
history based page dirty rate (HIST) which are accurate up to 90 percent of actual
results. They proposed a model called Activity Based Sector Synchronization
(ABSS) [22] which migrates virtual machine efficiently in a timely manner. The
Chapter 2. Aim, Objectives, Research Questions & Related Work 14
ABSS algorithm predicts the sectors that are likely to be altered. This algorithm
helps in minimizing the total migration time and the network bandwidth.
Robert et al. [23] worked on both disk related and network related issues.
They worked on migrating local persistent state by combining block level solution
with pre copy state. When the virtual machine need to migrate, disk state is first
pre-copied to the destination host. The virtual machine will still run during
this stage. After some time this mechanism starts XEN to migrate memory and
CPU states of the virtual machine. All the changes made to the disk on source
side during the migration process are recorded in the form of deltas (Unit which
contain written data, its location and size). These deltas are sent to destination
and applied to the image. This mechanism has downtime of 3 seconds in LAN and
68 seconds in WAN. They combined Dynamic DNS (DDNS) with IP tunneling
to retain old network connections after live migration. When the virtual machine
in source host goes to pause state, a tunnel is created using Linux iproute2. This
tunnel is between the virtual machine’s old IP address and the new IP address at
destination. This mechanism discards / drops the packets on source host’s side
during the final stage of migration. The packets related to the old connection are
forwarded from source host to IP address of the virtual machine at destination
via tunnel. The virtual machine is configured with a new dynamic IP address
to prevent it from using the old tunnel for new connections. Thus, the virtual
machine will now have two IP addresses. The disadvantage of this mechanism is
that the tunnel between source and destination will not be closed until the old
connections end. This requires the source host to run until old connections are
closed.
Kazushi et al. [13], used data duplication technique to propose a fast virtual
machine storage migration. This technique will reduce the volume and time of
data transfer. Suppose there is a situation where a virtual machine should migrate
frequently between site A and site B. Migrating a large disk between sites A and
B will waste a lot of network resources and will take a long time. These frequent
disk migrations can be eliminated by using duplication. When the VM is migrated
for the first time from site A to site B, full disk is copied to site B. When the VM
migrates back to site A, only changes made on disk in site B are replicated to
disk in site A. A new diff image structure, which is a new virtual machine image
format and Dirty Block Tracking (DBT) are developed to track the changes. This
technique successfully reduced the migration time from 10 minutes to 10 seconds.
Travostino F et al. discussed about the advantages and requirements of long-
haul live migration [5]. Their experiment proved that virtual machines can mi-
grate over WAN across long distance geographical locations instead of being lim-
ited to small local datacenters. Their scheme has a dedicated agent called as
Virtual Machine Traffic Controller (VMTC) that created dynamic tunnels be-
tween clients and the virtual hosts. The VMTC is responsible for migration of
the virtual machine and it maintains connectivity with the destination host where
Chapter 2. Aim, Objectives, Research Questions & Related Work 15
the virtual machine should be migrated. The VMTC communicates with Authen-
tication, Authorization and Accounting (AAA) module to get authentication in
the form of a token to setup an on-demand, end-to-end path. It is the responsibil-
ity of the VMTC to migrate the virtual machine and reconfigure the IP tunnels
so that it can seamlessly communicate with its external clients. When a virtual
host migrated from one host to other host the tunnel is broken and a new tunnel
is created between the client and virtual host. They migrated their virtual ma-
chine between Amsterdam and San Diego using a 1Gbps non-switched link with
Round Trip Time (RTT) of 198ms. The application downtime observed during
the migration is in the range of 0.8 to 1.6 seconds. Their experiment results show
that, when compared to LAN migration, the downtime of migration over WAN is
just 5-10 times higher, but the round trip time is more than 1000 times.This may
be due to the dynamically established link (light path) of 1Gbps using layer-1 and
layer-2 circuits without routing protocols. Anyhow, layer-3 protocols are used for
communication between VM and clients.
Eric et al. [14] used Mobile IPv6 (MIPv6), a protocol for managing mobility
of mobile nodes across wireless networks, to support live migration of virtual
machine over WAN. The virtual machines behave as mobile nodes by enabling
MIPv6 [24] in the TCP/IP protocol stack. This scheme enables a virtual machine
to retain its IP address after migrating from the source host. It eliminates the
need of DNS updates [23] on the virtual machine. The disadvantage in this scheme
is that the virtual machine’s kernel must be modified to support mobility (unless
that was already preconfigured).
Solomon et al. presented a live migration approach using Proxy Mobile IPv6
(PMIPv6) protocol [15]. This approach is similar to MIPv6 [14] but it does not
require installation of mobile software on a virtual machine [25]. However, it
requires specific infrastructure deployed in the network. In this experiment the
source host and destination host act as Mobile Access Gateways (MAGs) and are
connected to the Local Mobility Anchor (LMA). This choice in their experiment
was just for convenience, to keep the experiment testbed simple. The LMA and
MAG are entities independent infrastructure elements. The LMA and MAGs are
equivalent of a home agent and foreign agent in Mobile IP. The LMA arranges that
packets for the mobile node and forwarded them to the MAG where the node is
currently located. The MAG is responsible for emulating the mobile node’s home
network and for forwarding packets sent by the mobile node over the tunnel to
LMA. LMA acts as a central management node tracking the virtual machines.
When a virtual machine is booted on source machine, suppose MAG1 it registers
with LMA via Proxy Binding Unit (PBU) and Proxy Binding Acknowledgement
(PBA). Upon successful registration a tunnel is created between MAG1 and LMA.
Now the virtual machine can be connected to outer world via LMA. When the VM
is migrated to other location, for example toMAG2, the tunnel between MAG1
and LMA is broken and the VM is deregistered via PBU and PBA messages. Now
Chapter 2. Aim, Objectives, Research Questions & Related Work 16
the tunnel is created between MAG2 and LMA using the same previous process
that happened for MAG1. In this solution VM will retain its ip address and
network connections after the migration. Their experiment results showed that,
this approach migrated virtual machine with minimum downtime when compared
to MIPv6 approach. In this experiment authors used iSCSI shared storage to store
VM’s disk to avoid disk state migration.
Forsman et al. [16] worked on automatic seamless or live migration of virtual
machines in a data center. They developed an algorithm that migrates virtual
machine based on the CPU work load. They used two strategies called push
and pull. When the physical host’s workload crosses higher threshold level, the
hotspot is detected and the push phase initiates virtual machine migration to
other underutilized physical host to balance the load. When a physical host is
underutilized and is below the lower threshold level the pull phase is initiated
to balance the system by requesting virtual machines from other physical nodes.
They also used a concept called variable threshold, which varies the threshold
level of the load periodically, which resulted in a more balanced system. They
successfully simulated their technique using OmNet++ simulation software.
Chapter 3
XEN Hypervisor and Storage Solutions
In this chapter we described about the Xen hypervisor and various stor-
age solutions we used to analyse the disk performance during the live migration
of virtual machine. Xen, Kernel Virtual Machine (KVM), VMware ESXi and
Hyper-V are some of the widely used type-1 hypervisors. Among these hyper-
visors VMware ESXi and Hyper-V are expensive and proprietary software with
restrictive licence scheme. So, we restricted ourselves to open source hypervisors
like Xen and KVM. KVM requires the physical host to support hardware assisted
virtualization (HVM) but unfortunately our testbed doesn’t have hardware sup-
porting HVM. So, we used Xen hypervisor and created para-virtualized virtual
machines. We briefly described Xen hypervisor along with its features in section
3.1.
We used internet Small Computer Interface (iSCSI), Network File Storage
(NFS) and Distributer Replicated Block Device (DRBD) storage solutions to store
the disk of the virtual machines. These are the three widely used storage solutions
supported by various Hypervisors to store the disk state of virtual machines. Each
storage solution has its own advantages and disadvantages. A short comparison
of these three storage solutions is shown in table 3.1.
Among these three solutions DRBD and iSCSI are block storage solutions and
NFS is a file storage solution. In this thesis we used Xen hypervisor to migrate a
virtual machine while conducting disk I/O tests. Our virtual host has its disk at
shared or replicated storage solutions. So, the Xen migrates only the memory and
CPU states leaving the disk state. The performance of VM’s disk I/O operations
depends on underlying storage solutions. So, we assume that our results are valid
for other available hypervisors.
17
Chapter 3. XEN Hypervisor and Storage Solutions 18
one of the five1 type-1 or bare metal hypervisor that are available as open source
[42]. Xen allows multiple operating systems to run in parallel to host operating
system. Xen is used for different open source and commercial application such
as desktop virtualization, server virtualization, Infrastructure as a service (IaaS),
security, hardware and embedded appliances. Today Xen hypervisor is powering
large clouds in production [20]. The Xen hypervisor is responsible for handling
interrupts, scheduling CPU and managing memory for the virtual machines.
Dom0 or the Domain-0 is the domain in which Xen starts during the boot.
From the Xen architecture which is shown in figure 3.1 we can see that Dom0
is the privileged control domain which has direct access to the underlying hard-
ware. The Dom0 has the toolstack which is a user management interface to the
Xen hypervisor. Xen toolstack can create, manage and destroy virtual machines
or domUs which are unprivileged domains [20]. Xen supports hardware virtual-
ization and Paravirtualization. In hardware virtualization, unmodified operating
systems can be used for the virtual machines whereas Paravirtualization requires
modification to the operating system’s kernel running inside virtual machines.
Doing so will increase the performance of paravirtualized hosts. The host operat-
ing system should be Xen Paravirtualization enabled to create virtual machines.
The Linux kernels before 2.6.37 version are not Paravirtualization enabled. Their
kernels should be recompiled to enable Paravirtualization. All the Linux kernels
released after 2.6.37 version are by default Xen Paravirtualization enabled.
1
The other open source hypervisors are KVM, OpenVZ, VirtualBox and Lguest.
Chapter 3. XEN Hypervisor and Storage Solutions 19
Xen allows virtual machines to migrate between hosts while the guest operat-
ing system is running. This feature is called live migration. In Xen, the demons
running in the Dom0 of source and destination hosts takes the responsibility of
migration. The memory and CPU states of the virtual machine are migrated
from source machine to destination machine by the control domain. The Xen
Hypervisor copies memory pages in series of rounds using Dynamic Rate-limiting
and Rapid Page Dirtying [8] techniques to reduce the service downtime. Dy-
namic Rate-Limiting algorithm adapts bandwidth limit for each pre-copy round
and is used to decide when the pre-copy stage should end and stop-and-copy
phase should start. Rapid Page Dirtying algorithm is used to detect the rapidly
dirtied2 pages and skip them from copying them in pre-copy stage. Xen uses a
microkernel design which consists of small footprint and interface that is around
1MB of size making it more secure and robust than the other available hypervi-
sors. Xen hypervisor is capable to run main device driver for a system inside a
virtual machine. The virtual machine that contains main device driver and device
driver can be rebooted leaving the rest of the system unaffected [20]. This feature
of Xen is called driver isolation. Driver isolation is a safe execution environment
which protects the virtual machines from any buggy drivers [55]. Xen is operating
system agnostic which means different operating systems like NetBSD and Open
Solaris can be hosted.
Basically Xen supports two types of virtual Block Devices named ‘Phy’ and
‘file’. Phy is the physical block device which is available in the host environment
where as file is the disk image which is available in the form of a file in the host
computer. The loop block device is create from the available image file and the
block device is handled to the domU. Shared storage solutions like iSCSI use Phy
and NFS use file.
DRBD’s functions are purely implemented using Linux kernel module. Fea-
tures like flexibility and versatility makes DRBD suitable for providing duplica-
tion solution to many applications. DRBD package has two main components
namely, the driver code and the user tool. The driver code runs in kernel space
and the user tool is used to control the driver code and cluster management pro-
grams from user space. Drbdadm, drbdsetup, drbdmeta are some of the user tools
used to communicate with kernel module for managing and configuring the re-
sources of DRBD. DRBD remained an out-of-tree kernel module for many years
[43] but was then integrated into the Linux kernel from version 2.6.33 release.
We used DRBD v8.4.3 in our experiments. DRBD installation and configuration
details are described in the appendix.
DRBD has two operation modes namely single primary mode and dual pri-
mary mode. In single primary mode only one of the two disks in the cluster
acts as primary disk. All the changes done to the data on this primary disk are
replicated to the other peer disk which is at a different physical location. The
secondary disk can be used as backup in case of data loss on the primary disk.
The dual primary mode was first introduced in DRBD v8.0 and disabled by de-
fault. In dual primary mode both the disks in the cluster act as primary disks.
The dual primary mode can be integrated with Xen hypervisor to eliminate disk
migration during the live migration [2]. The changes we made in “/etc/drbd.conf ”
Chapter 3. XEN Hypervisor and Storage Solutions 21
configuration file enabled dual primary mode and integrated DRBD with XEN.
These configuration changes are described in the appendix.
Order of events from 1 to 4 is always the same. But the write event on local
disk can take place anytime independently of events 2 to 4.
Chapter 3. XEN Hypervisor and Storage Solutions 22
3.2.2.1 Protocol A
3.2.2.2 Protocol B
3.2.2.3 Protocol C
is exported from the target and it is seen as a local disk by the operating system
of the iSCSI initiator. The applications running on the initiator are not aware of
the iSCSI target. Here the initiator runs the local file system which reads from
and write the data blocks to the iSCSI target when required. This is displayed
in the figure 3.9. In Xen virtualization iSCSI is used to store the disks of virtual
machine so that the disks are available even after migrating virtual machine to
another physical host. This requires the two physical hosts to connect with the
iSCSI target. When the virtual machine is running on physical host A, it access
the disk which is at other location (iSCSI target) using iSCSI initiator drivers
running on physical host A. When it is migrated to the physical host B (which
is configured to connect the iSCSI target), it will resume using the disk from the
iSCSI target. The process to configure iSCSI to enable live migration using Xen
hypervisor is described in the appendix.
The physical node hosting a virtual machine needs an iSCSI initiator con-
nected to the network in order to access storage from the target. An iSCSI driver
can act as initiator with a standard network card or with a TCP offload engine
(TOE) network card which reduces the CPU utilization. The storage device must
also implement the iSCSI protocol stack on the target side. Unlike NFS which is
a file storage solution, block storage solutions like iSCSI doesn’t have file locking.
In iSCSI the target can’t connected to multiple initiators at same time unless
the target uses a cluster aware file systems like GFS and OCFS2. The multi-host
feature which is disabled by default should be enabled. In iSCSI each initiator has
a unique iSCSI Qualifier Name (IQN) id which is used to connect to the target.
Chapter 3. XEN Hypervisor and Storage Solutions 29
The iSCSI target generally grants access control based on the IQN. The virtual
machine can’t be created from different location using the same disk from iSCSI
target at a same time.
NFSv4 is the latest protocol in which different protocols like nfs3 ,mountd, nsm
and nlm [56] are integrated into a single protocol. Mountd protocol is responsible
for mounting exported system so that it can be accessed by nfs. Network Status
Monitor (nsm) protocol is used to monitor the machines state i.e., client or server.
Network Lock Manager (nlm) protocol manages a lock system which avoids the
clients to modify the data at same time.
We had used NFSv4 to create a NFS server-client setup in our laboratory
test-bed using three computers. We had one NFS server to share storage with
the two NFS clients. We created disk images for virtual machine and placed them
in a shared directory. If a virtual machine running on client 1 make changes to
the disk image the changes are written to the NFS server over the IP network.
The changes made on NFS server are made available to the NFS client 2. When
the virtual machine is migrated from client 1 to client 2 it uses the disk available
at client 2 and works as usual. NFS and Xen configurations to support live
migration are described in the appendix.
3
Nfs is the base protocol which allows to create, read, write and search files.
Chapter 4
Research Methodology and Test-bed
4.1.1 Stage 1
In this phase we conducted a thorough literature study to gain proper tech-
nical knowledge in the area of virtualization. We had conducted this study by
reading various articles, journals, books and web pages. We searched several
databases like IEEE explore, Inspec, Scopus, etc. to identify honest and quality
related articles.
31
Chapter 4. Research Methodology and Test-bed 32
4.1.2 Stage 2
In this phase we had conducted experiments to evaluate the disk performance
of virtual machine during the live migration over WAN or internet on laboratory
test-bed using various disk sharing solutions. There are three ways to achieve the
aim of our proposal i.e. mathematical modelling, real-time implementation and
simulation [18]. Mathematical modelling is very complex and not fully accurate
due to the simplifications done in modelling process while simulation is a simpli-
fied model of real time environment created to simulate real conditions virtually.
Simulations are used when the real time environment is unavailable. Different
simulators programs may give varying results. For example OPNET and Net-
work Simulator (NS-2), the popular network simulators gives different outputs
even if they were given same input parameters. So, users should be careful while
selecting the simulation software. To analyze the disk performance of a virtual
machine we required limited resources, for example 4 computers and a switch.
This made us to choose laboratory testbed to conduct our experiments. In cases
like ours, the laboratory experiment brings us closer to real-time environment
and gives advantage over simulation and mathematical analysis. Figure 4.1 is the
simplified experimental setup we used to collect the disk performance data.
4.1.3 Stage 3
In this stage we analyzed the performance of the three disk storage solutions
with the help of the observed performance metrics.
Chapter 4. Research Methodology and Test-bed 33
• Site 1 and Site 2 runs on HP dx5150 Small Form Factor with AMD Athlon(tm)
64 Processor 3200+, 1GB of RAM, 40GB Samsung SP0411C HDD and a
gigabit Ethernet adapter. These machines have Ubuntu 12.04 LTS Desktop
64 bit operating system.
• NFS and iSCSI servers runs on HP dx5150 Small Form Factor with AMD
Athlon(tm) 64 Processor 3200+, 2GB of RAM, 80GB Seagate HDD and a
gigabit Ethernet adapter. This machine has Ubuntu 13.10 64 bit operating
system.
• Virtual machine was allotted with 256MB of RAM, 4GB storage and 1
virtual CPU. This virtual machine has Ubuntu 12.04 LTS desktop 32-bit
operating system
Chapter 4. Research Methodology and Test-bed 34
• Client run on Lenovo Y480 notebook with 8GB of RAM, Intel core i7 CPU
and gigabit Ethernet NIC. This machine has Windows 8.1 64 bit operating
system.
• 16.0 X 10/100 D-Link 1016 DES unmanaged switch is used to connect Site
1, Site2, Storage server and client.
4.4.1 Iometer
Iometer [26] formerly known as “Galileo” is a very popular I/O subsystem
benchmarking tool developed by Intel Corporation. It was first released on 17th
February 1998. It can be used as both workload generator (Dynamo) and mea-
surement tool (Iometer) [26] on a single or networked systems.
The uses of Iometer are described below:
4.4.2 NetEm
Network emulator, in short NetEm [28] is a Linux kernel based program
that we used to emulate the properties of WAN over LAN. NetEm introduces
packet delay, packet loss, packet duplication and packet re-ordering. NetEm is
enables in all Linux distributions. It has two components, Command-line utility
and Kernel module [29]. The command line utility uses netlink socket interface
to communicate with kernel module. The NetEm command used to introduce
network delay is described in appendix.
4.4.3 Wireshark
Wireshark is an open source network packet analyzer which is used for cap-
ture, analyses of network packets. Wireshark is available on Linux, OS X, Win-
dows and many other operating systems. Some important purposes of wireshark
[31] are troubleshooting, examining network problems, learning network protocol
internals, etc.
4.5.1 Scenario 1
In this scenarios we conducted IO tests to analyze the disk performance of
virtual machine during the live migration over WAN using iSCSI shared storage.
In this scenario we have four computers in which two computers act as physical
hosts hosting virtual machines1 . These two physical hosts also act as iSCSI
initiators (iSCSI clients). One computer act as iSCSI target (server) and one
computer is used for collecting IO results using Iometer. This scenario is shown
in figure 4.2. Here we used NetEm to emulate WAN by introducing network delays
on the interfaces of the two physical hosts and iSCSI target. For example, we
introduced 15ms delay on the three interfaces so that the network delay between
the two physical hosts is 30ms and the delay between physical host and iSCSI
target is also 30ms.
1
The disk size of our virtual machines is 4GB
Chapter 4. Research Methodology and Test-bed 36
4.5.2 Scenario 2
In this scenarios we conducted IO tests to analyze the disk performance of
virtual machine during the live migration over WAN using NFS shared storage.
In this scenario we have four computers in which two computers act as physical
hosts hosting virtual machines. These two physical hosts also act as NFS clients.
One computer act as NFS server and one computer is used for collecting IO
results using Iometer. We followed the same procedure described in Scenario 1
to introduce network delay.
4.5.3 Scenario 3
In this scenarios we conducted IO tests to analyze the disk performance of
virtual machine during the live migration over WAN using DRBD storage. We
had three computers in this scenario, in which two computers act as physical
nodes hosting virtual machines. Here when the two physical hosts are booted,
the DRBD checks the status of the disks at both the sites. If the two disks are
up-to-date (synchronized) the connection is established between the two hosts. If
any of the two disks is not up-to-date, it is synchronized from the last updated
disk and the connection is established. The rate of synchronization is described in
chapter 3.2.5. When the connection is established between the two sites DRBD
asynchronous protocol C is executed on these two physical hosts to create a cluster
with the two sites. If a virtual machine makes changes to disk on site 1 the changes
are replicated to the disk on site 2. One computer is used for collecting IO results
using Iometer. We used NetEm to introduce delay on both hosts. We introduced
equal amount of delay on the interfaces of both sites to achieve required delay.
For example we introduced 45ms delay on both sites for 90ms delay.
According to [58], the E2E delay components are divided into four categories.
They are processing delays, transmission delays, propagation delays and queuing
delays. The processing delay is added by the intermediate devices which are in
the path between the two end points. It is the time taken by the intermediate
devices to process the data packet and retransmit it to next node or end point.
The processing delay is independent of network load but it depends on the node
hardware, processing or computing efficiency and the protocol stack used. The
processing delay time is a stochastic random variable as it is different for each
individual packet. Transmission delay is the time taken to transmit a packet
over the communication link and it depends on the link speed of the hardware.
Propagation delay is the time taken for the packet to travel from one end to
another end over a communication channel. It depends on the type of channel
used between the endpoints. For example the propagation time taken on optical
fiber is different to the propagation time taken on satellite link. While coming to
the queuing delay, it is stochastic in nature and depends on the packet’s waiting
time in the buffer of the network devices before being processed or serviced.
From [58], most of the End to End delay distributions observed over WAN
are heavy-tailed distributions with gamma shape (84%). While the other ob-
served delay distributions are, gamma-like with Gaussian or triangle lob distri-
butions (6.3%), two gamma-like shaped distributions (2.8%), distribution with
many peeks (5%) and other distribution classes (1.5%). The network delays with
non-normal or heavy tail distributions degrade the network performance due to
their high probability spread in the values far from mean delay. Since, the heavy
tail distribution spread their probability curve to a higher delay time’s, a higher
waiting time is probable, this may increases the waiting times drastically at some
point of time. The impact of heavy-tailed service time destitution on the sys-
tem performance are discussed in [60]. Due to heavy-tailed distributed delays in
networks, the system takes more time to reach steady state. Since, the transient
nature of heavy tail distributions are high the waiting time distributions can’t
be accurately estimated. So the E2E heavy-tail delay distributions are difficult
to emulate. So, for simplicity most of the WAN emulators use Normal or Pois-
son distributions. These Normal / Poisson distributions decrease the probability
values exponentially on either side of mean, which makes it easy to emulate the
E2E delay characteristics of WAN. In our thesis we emulated the properties of
WAN using NetEm tool which uses normal distribution to introduce network de-
lay on the interface of the host. We introduced normally distributed intra-Europe
(30ms), trans-Atlantic (90ms) and trans-Pacific (160ms) delays (example shown
in appendix) to analyze the performance of virtual machine’s disk while migrating
over WAN. The network latencies we used in our thesis are observed from the IP
latency statistics published by Verizon Enterprise Solutions [49].
Chapter 5
Experiment Results
In this chapter we discuss the performance of iSCSI, NFS and DRBD storage
solutions under the influence different network delays. We had emulated a WAN
by introducing various network delays using NetEm. We had conducted a total
of twelve experiments for each of the three disk storage solutions by varying four
network delays. We ran every experiment for 40 iterations and calculated confi-
dence intervals for the mean values to deal with the stochastic variations caused
by I/O interrupts, kernel process delays, IP timers and measurement errors.
We used the selected performance metrics discussed in chapter 4 to analyze the
performance of iSCSI, NFS and DRBD by varying the network delays. First we
conducted experiment without introducing any network delay. This is a baseline
scenario that represents a campus network. Then we introduced 30ms RRT net-
work delay to emulate an inter-European WAN. Next we introduced 90ms RTT
network delay to emulate Trans-Atlantic WAN. Finally we conducted experiments
by introducing 160ms RTT network delay to emulate a trans-Pacific WAN.The
chosen values are the average latency guarantees, expressed as round-trip times
(RTTs), specified in the SLA for networks operated by Verizon Enterprise Solu-
tions [49].
We calculated 95 % CI [50] for the sample mean of the performance metrics
observed during different network delays. We compared the performance of the
three storage solutions with respect to performance metric and network delays.
The confidence interval is calculated using the formula shown in (1).
s
x̄ ± t(n−1)(1−α/2) √ (5.1)
n
40
Chapter 5. Experiment Results 41
With 95% confidence we can say that I/Os handles by iSCSI lies in the range
of 21.68 to 22.72 with standard deviation of 2.12 and relative error of 2.34%. The
I/O handled by NFS lies in the range of 49.25 to 52.44 with standard deviation
of 6.50 and relative error of 3.13%. I/O’s handled by DRBD lies in the range of
52.37 to 54.17 with the standard deviation of 3.46 and the relative error of 1.69%.
With 95% confidence we can say that the I/O operations of iSCSI operations
lies between 14.77 and 16.01 with standard deviation of 2.37 and relative error of
4.01%. The I/O operations handled by NFS lied in the range of 36.41 to 38.45
with standard deviation of 4.14 and relative error of 2.71%. The I/O operations
handled by DRBD are in range of 25.83 and 26.64 with standard deviation of
1.55 and relative error of 1.54%. When comparing block storage solutions DRBD
outperformed iSCSI. The I/O operations handles be DRBD are 70.50 % more
than iSCSI.
With 95% confidence we can say that the I/O operations of iSCSI lies in
the range of 9.94 to 10.59 with standard deviation of 1.33 and relative error of
3.16%. The I/O operations handled by NFS lies in the range of 23.76 to 25.17
with standard deviation of 2.88. The I/O operations handled by DRBD lies in
the range of 15.38 to 16.03 with standard deviation of 1.24 and relative error of
2.06%. When file and block storages solutions are compared, NFS has better
performance than the two block storage solutions. When block storage solutions
Chapter 5. Experiment Results 44
When a virtual machine was migrated over Inter-Europe network using differ-
ent storage solutions, DRBD outperformed iSCSI and NFS. The average volume
of data read and written per second on DRBD disk is 0.208 MB. When coming
to iSCSI and NFS, 0.085 and 0.197 MB of data is read or written per second
on the disks. The performance of iSCSI is 144.70% less than the performance of
Chapter 5. Experiment Results 46
DRBD whereas the performance of NFS is only 5.58% lower. With 95% confi-
dence we can say that the performance of iSCSI lies in the range of 0.083 to 0.087
with standard deviation of 0.008 and relative error of 2.5%. The 95% confidence
interval values of NFS lies in between 0.190 to 0.204 with standard deviation of
0.026 and relative error of 3.45%. With 95 % confidence we can say that the
performance of DRBD lies in between 0.2045 to 0.211 with standard deviation
of 0.013 and relative error of 1.69. When block and file storage solutions are
compared, DRBD slightly outperformed NFS. When two block storage solutions
are compared, DRBD outperformed iSCSI by 142.28%.
The 95% confidence values along with standard deviation and relative error
of MBPS read and written on the disks of iSCSI, NFS and DRBD are shown in
the table 5.8. The relative errors of iSCSI, NFS and DRBD are 3.56%, 3.09%
and 2.06% respectively. When block and file storage solutions are compared, NFS
outperformed both iSCSI and DRBD. When block storage solutions are compared
DRBD outperformed iSCSI by 52.5%.
From the below graph we can see that NFS outperformed the other storage
solutions when a virtual machine is migrated over campus, trans-Atlantic and
trans-Pacific network. In inter-Europe network DRBD outperformed NFS with a
small margin. NFS network storage solution made best use of its efficient caching
technique to handle more megabytes of data when compared to iSCSI and DRBD.
and 2101.92 milliseconds. The performance of DRBD is 238.71% greater than the
performance of NFS shared storage and 214.64% greater than the performance
of iSCSI. The 95% confidence interval of the sample mean MRT value from 40
iterations along with standard deviation and relative error are shown in table
5.10. The relative error values of iSCSI, NFS and DRBD are 4.15%, 4.27% and
2.83% respectively. When block and file storage solutions are compared DRBD
outperformed NFS.
chine is migrating over campus network, the iSCSI storage has least maximum
response time, while DRBD outperformed other storage solutions when virtual
machine is migrated over inter-Europe, trans-Atlantic and trans-Pacific networks.
Among the three storage solutions iSCSI and NFS are shared storage solutions
where the disk is located at another location. The I/O operations generated from
the virtual machine are encapsulated and forwarded to server using IP packets.
But DRBD is a disk replication solution where the disk resides at the client and
the I/O changes mode on the disk are replicated to clustered disk using IP pack-
ets. When a VM is migrated over WAN, the MRT value observed in iSCSI and
NFS storages system is higher when compared to DRBD because iSCSI and NFS
have a storage server which is located at a remote location and the time taken by
the I/O packets to reach the server adds up to the observed I/O response time.
When coming to DRBD, there is no third device between the two clustered disks
to add additional travel and process times. This will also reduce the network con-
sumption when compared to iSCSI and NFS. So the DRBD outperforms iSCSI
and NFS storage solutions when migrating over WAN.
estimated using the formula described in chapter 3.2.5. In this experiment both
the DRBD nodes are up-to-date which eliminated the initial disk synchronization
phase. The iSCSI consumed 74.23% more network resources than DRBD whereas
NFS consumed 99.81% more network resources. DRBD uses less network band-
width than other two solutions because, in DRBD only the changes made by the
write I/O operations are replicated to peer disk over the network. Whereas in
iSCSI and NFS both write I/O operations and read1 I/O operations uses network
to reach the storage server. So, using DRBD along with Xen when migrating
virtual machine over low latency WAN reduces the network bandwidth.
1
Actually not all read operations. This is because some read I/O‘s can access cached data.
Chapter 6
Conclusion and Future work
6.1 Conclusion
In this thesis we had analyzed the disk performance of virtual machines during
live migration over networks with different varying network latencies and using
different storage solutions. We had used iSCSI, NFS and DRBD storage solutions
to store the disk of the virtual machine and analyzed disk I/O performance of
a simulated webserver with 80% reads and 20% writes using Iometer disk ana-
lyzer tool. We had conducted our experiments by varying three network delays
to emulate WAN. When the disk performance of virtual machine using different
storage solutions is compared with varying network delays, the experiments con-
ducted with NFS storage solution handles more I/O operations and fetched more
megabytes of data when compared to iSCSI and DRBD storage solutions. The
performance of DRBD is slightly inferior to the performance of NFS and clearly
out performed iSCSI storage solutions. In NFS, it’s efficient caching mechanism
caches the most frequently read operations and reduces the need of accessing the
disk at another location over a network. This caching technique could be the
main reason for the better performance of NFS storage solution when the VM
hosts a server handling more read operations.
In case of maximum response time (MRT) metric, DRBD outperformed iSCSI
and NFS storage solutions. This performance metric is the total time taken
between the initiation and completion of an I/O operation. The iSCSI and NFS
are shared storage solutions where the virtual machine’s disk is stored at a remote
location and read and write operations are carried to the disk over IP protocol
using the underlying WAN. The time taken by the read and write operations to
reach the remote disk will add up to the I/O response time and thereby increasing
it. On the other hand, DRBD it is a disk replication technique where the virtual
machines disk is stored in the local disk and only disk changes are replicated
to other peer disk. In DRBD the I/O operations are read and written to the
local disk. The network overhead consumed by DRBD is smaller than that used
by iSCSI or DRBD as only disk modifications are replicated to other peer node.
Among all our performance metrics MRT is crucial because if it is high, the
53
Chapter 6. Conclusion and Future work 54
client accessing the services from the virtual machine may detect the performance
difference. This may lead to poor Quality of Experience (QoE). Consequently, we
recommend DRBD storage solution for storing the virtual machine’s disk during
live migration over WAN.
From our experiment results we can say that even though NFS handles more
I/O operations and throughput, from the IO tests with 80% read 20% write
(which is a typical workload for a webserver) it had recorded more MRT and
network bandwidths effecting the VM’s performance during the live migration.
Suppose if the VM hosts a server which need to handle more write operations
than a webserver (for example in the case of a gaming server) the performance
of NFS shared storage may be further decreased due to high MRT and network
bandwidth. From our results we can conclude that DRBD is best suitable for
storing disk of Virtual machines during the live migration of VM over WAN and
iSCSI could be the next best option.
In this appendix we describe the procedure to install various tools and soft-
ware we used in our thesis. This appendix is divided in three sections. In section
one we described the procedure to install and configure NFS, iSCSI and DRBD.
The procedure to install and configure Xen hypervisor to support various storage
solutions is described in section two. Finally other tools like Iometer and NetEm
were discussed in section three. All the shell commands described in this ap-
pendix requires root privileges and we assume that all the devices are connected
to internet.
I. Section A
In this thesis we analysed the performance of iSCSI, NFS and DRBD storage
systems. Among these three solutions iSCSI and NFS are shared storage systems
whereas DRBD is replicated storage system.
56
Appendix 57
a. NFS Server:
The system should have root access to install and configure NFS
on Linux. We have to install the NFS software using apt-get install
command.
mkdir /vm1
chow nobody:nogroup /vm1
In the above command rw indicates that the client can read and
write to the files in the shared directory and the command sync indi-
cates that the changes made by the client are sent synchronously to the
server. In synchronous mode, the write requests issued from the client
side are considered complete after the changes are confirmed on the
server side. In asynchronous mode the write event issued on server side
are considered complete after they reaches the server. When compared
to asynchronous mode, synchronous mode is safe and reliable.
We need to export the settings using the below command to make
them effective.
$ exportfs -a
$ exportfs -v
$ mkdir p /mnt/nfs/vm1
$ mkdir p /mnt/nfs/vm1
$ mount $ df -h
We can test the connectivity by creating a file from one of the client
and view it on other client and server. The virtual machine’s image
places in the server can be accesses by the virtual machine running at
client1 or client2.
ISCSITARGET ENABLE=true
Next, open the file “/etc/iet/ietd.conf ’ ’ and add the below lines in the
file by removing all the existing content.
Appendix 60
Target iqn.2014-04.com.example:shared.lun1
IncomingUser
OutgoingUser
Lun 0 Path=/dev/vg0/shared,Type=fileio
Alias LUN1
iqn.2014-04.com.example:shared.lun1 192.168.137.0/24
After saving the modified files the iSCSI target service should be
restarted.
$/etc/init.d/iscsitarget start
$node.startup = automatic
/etc/init.d/open-iscsi restart
We have two logical disks for the virtual machine. In drbd the disk is
called as resource. Let us suppose they are drbdvm-disk and drbdvm-swap.
In our experiment we used DRBD protocol C (synchronous) to replicate
disks. We followed the procedure in [36] and [35] to configure the DRBD.
The drbd configuration file “/etc/drbd.conf should be edited to establish a
connection between the two hosts.
include ”/etc/drbd.d/global common.conf”;
include ”/etc/drbd.d/*.res”;
# VM disk image
resource drbdvm-disk {
protocol C;
net {
allow-two-primaries yes; #Enables dual primary mode
}
syncer {
rate 500M;
verify-alg sha1;
}
on site1-HP-dx5150-SFF-PE680AV {
address 192.168.137.146:7789; #Host1
Appendix 62
device /dev/drbd1;
disk /dev/vg0/drbdvm-disk; #disk location on host1
meta-disk /dev/vg0/meta[0];
}
on site2-HP-dx5150-SFF-PE680AV {
address 192.168.137.122:7789; # Host2
device /dev/drbd1;
disk /dev/vg0/drbdvm-disk; #Disk location on host2
meta-disk /dev/vg0/meta[0];
}
}
# VM swap image
resource drbdvm-swap {
protocol c;
net {
allow-two-primaries yes; #Enables dual primary mode
}
syncer {
rate 500M;
verify-alg sha1;
}
on site1-HP-dx5150-SFF-PE680AV {
address 192.168.137.146:7790; #Host1
device /dev/drbd2;
disk /dev/vg0/drbdvm-swap; #Disk location on host1
meta-disk /dev/vg0/meta[1];
}
on site2-HP-dx5150-SFF-PE680AV {
address 192.168.137.122:7790; #Host2
device /dev/drbd2;
disk /dev/vg0/drbdvm-swap; #Disk location on host2
meta-disk /dev/vg0/meta[1];
}
}
the DRBD resources should be create on both the locations by using the
below commands.
$/etc/init.d/drbd start
The DRBD disk status can be seen by using the below commands. In
this stage disks on the both locations will be in inconsistent state.
$/etc/init.d/drbd status
$cat /proc/drbd
Next disks in both locations should be attached be using the below shell
command on both the nodes.
$drbdadm up all
Next the disks on node one should be made primary followed by second
node. This will synchronize both disks. This will take time depending on
network bandwidth and disk size. In LAN environment synchronization
may take upto 5 minutes.
The synchronization time left and disk status can be seen by using the
below shell command.
$cat /proc/drbd
Appendix 64
II. Section B
In this section we described the procedure we followed to install and configure
XEN Hypervisor.
After installing Xen hypervisor modify the grub to make Xen default
while booting.
Then reboot the system so that the system boots with Xen hypervisor.
After rebooting you can verify xen by typing xm list which displayes dom0
as result.
$reboot
$xm list
2. Network Configuration
We used Linux Bridge to emulate the network link between physical
host and virtual machine. The Ubuntu 12.04 LTS desktop version will have
preinstalled bridge-utils package or it can be installed using the below shell
command.
Appendix 65
auto xenbr0
iface xenbr0 inet dhcp # You can replace dhcp with manual
bridge ports eth0
auto eth0
iface eth0 inet manual
$/etc/init.d/networking restart
Or
$service networking restart
Note: You can also use synaptic package manager to install Xen hyper-
visor and network bridge [39].
$ wget http://ftp.se.debian.org/debian/pool/main/x/xen-tools/xen-
tools 4.3.1-1 all.deb
$dpkg i xen-tools 4.3.1-1 all.deb
name = ’vmtest’
dhcp = ’dhcp’
vif = [ ’mac=00:16:3E:C5:24:63,bridge=xenbr0’ ]
on poweroff = ’destroy’
on reboot = ’restart’
on crash = ’restart’
dd if=/dev/vg0/vmtest-disk of=/mnt/nfs/img/nfsvm-disk.img
dd if=/dev/vg0/vmtest-swap of=/mnt/nfs/img/nfsvm-swap.img
Appendix 68
III. Section C
In this section we described the procedure to download and configure miscel-
laneous tools like Iometer and NetEm.
$wget http://prdownloads.sourceforge.net/iometer/iometer-
2006 07 27.linux.i386-bin.tgz
2. Configuring NetEm
NetEm is a Linux based tool used to emulate the properties of WAN over
LAN by introducing network delay. Since the network delay is not uniform
in WAN, the delay variation can be normally distributed using NetEm. The
below shell command is used to introduce the mean trans-Atlantic delay of
90ms with a variation upto 10ms which is normally distributed. Users
can also model their own distributions my making changes to the iproute2
compilation code which is located at “usr/lib/tc”.
$tc qdisc change dev xenbr0 root netem delay 90ms 10ms distribution
normal
References
References
71
12. Yingwei Luo, Binbin Zhang; Xiaolin Wang; Zhenlin Wang; Yifeng Sun; Haogang
Chen, "Live and incremental whole-system migration of virtual machines using block-
bitmap", 2008 IEEE International Conference on Cluster Computing (CLUSTER), p
99-106, 2008.
13. Kazushi Takahashi, Koichi Sasada, Takahiro Hirofuchi, "A Fast Virtual Machine
Storage Migration Technique Using Data Deduplication", The Third International
Conference on Cloud Computing, GRIDs and Virtualization, 2012.
14. Eric Harney, Sebastien Goasguen, Jim Martin, Mike Murphy, Mike Westall, "The
efficacy of live virtual machine migrations over the internet", Proceedings of the 2nd
international workshop on Virtualization technology in distributed computing, 2007.
15. S. Kassahun, A. Demessie, and D. Ilie, "A PMIPv6 Approach to Maintain Network
Connectivity During VM Live Migration Over the Internet", in Proceedings of IEEE
CloudNet, Luxembourg, October 2014.
16. Mattias Forsman, Andreas Glad, "Automated Live Migration of Virtual Machines",
M.S.Thesis, School of Computing, Blekinge Institute of Technology, Karlskrona,
Sweden, 2013.
17. "Applied Research", [online]. Available:
http://en.wikipedia.org/wiki/Applied_research, [Accessed: 07/12/2013].
18. Guanhua Yan, Yuehui Jin, Ilulu Ma, Shiduan Cheng, Jian Ma, "AN EFFICIENT
METHOD TO SIMULALATE COMMUNICATION NETWORKS", International
Conference on Communications, June 2000, pp. 192-194.
19. Sherif Akoush, Ripduman Sohan, Andrew Rice, Andrew W. Moore and Andy Hopper,
“Predicting the Performance of Virtual Machine Migration”, IEEE International
Symposium on Modeling, Analysis & Simulation of Computer and Telecommunication
Systems (MASCOTS), pp. 37-46, Aug 2010.
20. “Xen Project Software Overview”, [Online]. Available:
http://www.wiki.xen.org/wiki/Xen_Project_Software_Overview, [Accessed:
07/12/2013].
21. Timothy Wood, Prashant Shenoy, Arun Venkataramani, and Mazin Yousif, "Black-box
and gray-box strategies for virtual machine migration", 4th USENIX proceedings,
conference on Networked systems design & implementation, April 2007, pp. 17-17.
22. Sherif Akoush, Ripduman Sohan, Bogdan Roman, Andrew Rice and Andy Hopper
, "Activity Based Sector Synchronisation: Efficient Transfer of Disk-State for WAN
Live Migration", IEEE 19th International Symposium on Modeling, Analysis &
Simulation of Computer and Telecommunication Systems (MASCOTS), Singapore,
July 2011, pp 22-31.
23. Robert Bradford, Evangelos Kotsovinos, Anja Feldmann and Harald Schioberg, "Live
wide-area migration of virtual machines including local persistent state", Proceedings
of the 3rd international conference on Virtual execution environments, June 2006, pp.
169-179.
72
24. D. Johnson, C. Perkins, and J. Arkko, “Mobility Support in IPv6,” RFC 3775, Jun.
2004.
25. L. Dunbar and B. Sarikaya, “Virtual Machine Mobility Protocol Using Distributed
Proxy Mobile IPv6,” IETF Internet draft, Feb. 2013, [Online]. Available:
http://tools.ietf.org/html/draft-sarikaya-nvo3-vmm-dmm-pmip-00/ [Accessed:
02/01/2014]
26. Wayne Allen, Ming Zhang, Vedran Degoricija and Daniel Scheibli, “Iometer user
guide”, Iometer project, September 2004.
27. Vince Westin, “User guide: iorate Open Systems I/O Driver and Measurement Tool”,
EMC corporation, 2005.
28. Linux foundation, “NetEm”, November 2009, [Online].
http://www.linuxfoundation.org/collaborate/workgroups/networking/netem,
[Accessed: 12/04/2014]
29. Pradyumna Dash, "Bandwidth Throttling with NetEM Network Emulation", June 2012,
[Online]. http://www.opensourceforu.com/2012/06/bandwidth-throttling-netem-
network-emulation/, [Accessed: 12/04/2014]
30. Greg Porter, Howto:Iometer, November, 2010, [Online].
http://greg.porter.name/wiki/HowTo:iometer [Accessed: 03/12/2013]
31. Ulf Lamping, Richard Sharpe and Ed Warnicke, "Wireshark User's Guide", [Online].
https://www.wireshark.org/download/docs/user-guide-a4.pdf, [Accessed: 30/06/2014].
32. George Crump, "What is NFS Caching?", June 2011, [Online] http://www.storage-
switzerland.com/Articles/Entries/2011/6/22_What_is_NFS_Caching.html [Accessed:
19/07/2014]
33. Prabhakar Chaganti, “Xen virtualization: A practical handbook”, 1st edition, Packt
Publishing, Birmingham, UK, December 2007.
34. Falko, "Xen Live Migration Of An LVM-Based Virtual Machine With iSCSI On
Debian Lenny", April 2009, [Online]. http://www.howtoforge.com/xen-live-migration-
of-an-lvm-based-virtual-machine-with-iscsi-on-debian-lenny/ [Accessed: 06/02/2014]
35. Keith Herron, "Live Migration and Synchronous Replicated Storage with Xen, DRBD
and LVM", April, 2010. [Online]. http://backdrift.org/live-migration-and-synchronous-
replicated-storage-with-xen-drbd-and-lvm/ [Accessed: 12/01/2014].
36. Xen Cluster: How to, February, 201, [Online] http://www.asplund.nu/xencluster/xen-
cluster-howto.html, [Accessed: 13/04/2014].
37. Ubuntu Xen Community, "XenProposed", September, 2012, [Online].
https://help.ubuntu.com/community/XenProposed, [Accessed: 23/12/2013].
38. Ian Jackson, "xen-tools – a straightforward VM provisioning/installation tool", August
2012, [Online]. http://blog.xen.org/index.php/2012/08/31/xen-tools-a-straightforward-
vm-provisioninginstallation-tool/ [Accessed: 21/01/2014]
39. Mehmet Batuhan Özcan, Gabriel Iro, “Paravirtualization Implementation In Ubuntu
With Xen Hypervisor”, B.Sc thesis, Dept. of Communication, Blekinge Institute of
Technology, Sweden, 2014.
73
40. Paul Virijevich, "Live migration of Xen domains", July, 2005, [Online]
http://archive09.linux.com/feature/55773/ [Accessed: 20/03/2014]
41. Falko, "Xen: How to Convert An Image-Based Guest To An LVM-Based Guest", April
2009, [Online], http://www.howtoforge.com/xen-how-to-convert-an-image-based-
guest-to-an-lvm-based-guest/, [Accessed: 20/04/2014]
42. Rodney Gedda, "5 open source virtualisation technologies to watch: Third-party
management tools growing around OS hypervisors", January 2011, [Online].
http://www.cio.com.au/article/373314/5_open_source_virtualisation_technologies_wa
tch/ [Accessed: 20/07/2014]
43. P. Reisner, F. Haas and L. Ellenberg, “The DRBD User’s Guide,” LINBIT Solutions
GmbH, 2011.
44. Peter Radkov, Li Yin, Pawan Goyal, Prasenjit Sarkar and Prashant Shenoy, "A
Performance Comparison of NFS and iSCSI for IP-Networked Storage", 3rd USENIX
Proceedings, Conference on File and Storage Technologies of USENIX Association,
2004, pp 101-114.
45. Ibrahim Yakti and Walid A. Salameh, "iSCSI (Internet Small Computer System
Interface) for your Startup", International Journal of Information and Communication
Technology Research, Vol 2, March 2012, pp. 294-303.
46. J. Satran, K. Meth, C. Sapuntzakis, M. Chadalapaka and E. Zeidner, " Internet Small
Computer Systems Interface (iSCSI)", RFC 3720, April 2004. [Online]
http://www.ietf.org/rfc/rfc3720.txt/ [Accessed: 23/07/2014]
47. White paper, "iSCSI Basics: A Practical Introduction", Mosaictec Solutions, 2006.
[Online] http://www.mosaictec.com/pdf-docs/whitepapers/iSCSI_Guide.pdf/
[Accessed: 09/06/2014]
48. S. Shepler, B. Callaghan, D. Robinson, R. Thurlow, C. Beame, M. Eisler and D.
Noveck, "Network File System (NFS) version 4 Protocol", RFC number 3530, April
2003. [Online] https://www.ietf.org/rfc/rfc3530.txt/ [Accessed: 23/06/2014]
49. Verizon Enterprise Solutions, "IP Latency Statistics", [Online].
http://www.verizonenterprise.com/about/network/latency/ [Accessed: 16/04/2014]
50. R. Jain, "The Art of Computer Systems Performance Analysis- Techniques for
Experimental Design, Measurement, Simulation, and Modeling," Wiley- Interscience,
New York, NY, April 1991.
51. David Marshall, “Top 10 benefits of server virtualization”, Virtualization report,
[Online]. http://www.infoworld.com/article/2621446/server-virtualization/top-10-
benefits-of-server-virtualization.html [Accessed: 16/09/2014]
52. Mohamed Fawzi, "Virtualization and Protection Rings (Welcome to Ring -1) Part I",
[Online], http://fawzi.wordpress.com/2009/05/24/virtualization-and-protection-rings-
welcome-to-ring-1-part-i/ [Accessed: 12/09/2014]
53. "Protection ring", [Online] Available: http://en.wikipedia.org/wiki/Protection_ring
[Accessed: 11/09/2014]
74
54. Gregory Steulet, "Dolphin Express Interconnect", [online],
http://www.trivadis.com/uploads/tx_cabagdownloadarea/DolphinInterconnect_U5.pdf
[Accessed:16/09/2014]
55. Jose Renato Santos, G. Janakiraman, Yoshio Turner, Ian Pratt, "Bridging the gap
between software and hardware techniques for I/O virtualization", USENIX 2008
Annual Technical Conference on Annual Technical Conference, p.29-42, June 22-27,
2008.
56. Frédéric Raynal, "NFS - Network File System", [online] Available:
http://www.linuxfocus.org/English/November2000/article164.meta.shtml [Accessed:
18/09/2014].
57. Paxson V, Floyd, Sally, "Wide area traffic: the failure of Poisson modeling",
IEEE/ACM Transactions on Networking, vol.3, no.3, pp.226-244, Jun 1995.
58. C. J. Bovy, H. T. Mertodimedjo, G. Hooghiemstra, H. Uijterwaal and P. van Mieghem,
“Analysis of end-to-end delay measurements in Internet”, in Proc. Passive and Active
Measurement Workshop (PAM 2002), Fort Collins, CO, Mar. 2002.
59. Wang, Junfeng and Zhou, Mingtian and Li, Yuxia: “Survey on the End-to-End Internet
Delay Measurements”, in Mammeri, Zoubir and Lorenz, Pascal (Edt): “High Speed
Networks and Multimedia Communications”, Lecture Notes in Computer Science,
Vol.3079, Springer Berlin Heidelberg, 2004, pp. 155-166.
60. Ward Whitt "The Impact of a Heavy-Tailed Service-Time Distribution Upon the
M/GI/s Waiting-Time Distribution", Queueing Systems, vol. 36, November 2000, pp.
71-87.
75