Data Centre Network Optimization
Data Centre Network Optimization
Received: 11 Jul 2023; Received in revised form: 15 Aug 2023; Accepted: 25 Aug 2023; Available online: 03 Sep Aug 2023
Abstract— Technology has advanced quite quickly in recent years, and IT and telecom infrastructure is
constantly growing. Due to an increased importance and emphasis of cloud computing many growing and
emerging enterprises have shifted their computing needs and wants ultimately to the cloud, which leads to
increase in inter-server data trafficking and also the bandwidth required for Data Centre Networking
(DCCN). Actually this multi-tier hierarchical architecture used in modern data centres is based on
traditional Ethernet/fabric switches. Researchers goal was to improve the data centre communication
network design such that the majority of its problems can be solved while still using the existing network
infrastructure and with lower capital expenditures. This is achieved through the deployment of OpenFlow
(OF) switches and the Microelectromechanical Systems (MEMS) based all Optical Switching (MAOS). This
will be beneficial in decreasing network latency, power consumption, and CAPEX costs in addition to helping
with scalability concerns, traffic management, and congestion control. Additionally, by implementing new
virtualization techniques, we may enhance DCCN's resource consumption and cable problems. In order to
address data centre challenges, the researcher came up with an entirely new novel flat data centre
coordination network architecture which is named as “Hybrid Flow based Packet filtering” (HFPFMAOS),
forwarding, and MEMS which is entirely based on optical switching, and will be finally controlled by
Software Defined Network (SDN) controller.
Keywords— Data Centre, Network Optimization, Software defined Networking, OpenFlow
• Cloud Based
2.3.1 Private
It is the on-site hardware that offers computing services and
stores data within a local network that is managed by an
organization's IT department. These are owned and run by
small private/public businesses, governmental
organization’s etc.
2.3.2 Cloud Based
Another name used for cloud data centres is co-
Fig. 2.3 DC Tiers
location/managed services provider. Co-located Data
Centers were created and maintained to offer specialized
infrastructure and provide various services to external 2.6 Design Factors for Data Centre Network
sources and parties. Major elements that force businesses to
• The following factors must be taken into account
either go for cloud computing or develop their own data
when planning and deploying DCCN architecture
centers are:
[3].
• Business market requirements
• Failure Impact & Application Resilience: All local
• Data privacy issues
and distant users will be unable to access their apps
• The increasing cost of related if DCCN fails.
equipment/infrastructure
• • Connectivity between the server and host:
2.4 Data Centre Communication Network (DCCN)
Servers must be linked together using several
Data services and data transit from the server to clients or redundant Ethernet connections.
other servers are the major goals of DCCN. These qualities
• Traffic Direction: In DCCN, the majority of traffic
are now required from the perspectives of dependability,
is between servers inside the data centre.
expansion, and efficiency for data centre communication
networks. • Agility: This term refers to the capacity for
assigning any service or application to any server
• Availability
at any moment in the data centre network while
• Scalability maintaining sufficient performance isolation and
• Flexibility security amongst various applications.
flows is running short, the incoming rate from 2.8 DCCN Topologies
numerous transmitters surpasses the rate at which it The architecture of a data centre communication network
can handle packet flows. This congestion causes the (DCCN), which serves as the foundation for its
delay time to lengthen and the receiver to begin applications, must be scalable, reliable, latency-free, and
dropping packets, which has an impact on various have adequate capacity to prevent network congestion.
applications, particularly those that depend on These qualities heavily rely on the network architecture in
latency, and lowers the network's overall which the DCCN is installed and are crucial to the overall
performance. effectiveness of the data centre. Figure below depicts the
• Managing the read/write, backup, and replication Common DCCN Architecture, as stated by Cisco [9].
traffic that moves within the data centre due to the
separation of application servers, databases, and
storage, which accounts for over 75% of all traffic
(i.e., east-west traffic between racks). Additionally,
because jobs are distributed across numerous servers
and assigned in parallel processing, DCCN's internal
traffic is increased.
• Support for a range of traffic patterns, such as long-
lasting, high-bandwidth "elephant" flows and short-
lasting persistent "mouse" flows. Examples of
entirely large data producers include particle
accelerators, planes, trains, metros, undergrounds,
self-driving cars, patient data in healthcare centers,
etc. In the year 2014, the Boeing 787 produces 40 Fig 2.4 Common DCCN Architecture
terabytes of onboard data every hour, of which 0.5
terabytes per hour is transmitted to a data centre. The
majority of flows within a data centre are typically Layered architecture has the advantage of enhancing the
mouse flows, while only a few elephant flows flexibility, modularity, and resilience of networks. Every
contain the majority of the data. Managing every sort layer in this design performs a separate role for unique
of traffic at once while keeping the overall network profiles, such as the routers in the core layer (i.e., the
delay within bounds is therefore a major issue. aggregation router and the border router), which use routing
• As a network grows, difficulties with scalability, protocols to decide how to forward traffic for both ingress
agility, and efficient resource use arise. In addition to and egress. Core switches, which offer incredibly flexible,
making fault handling and debugging more scalable connections with numerous aggregation layer
challenging as the number of communication devices switches, make up the Layer 2 domain. This layer serves as
increases, this will also result in an increase in a default gateway and a focal point for server IP subnets.
management overhead bytes on the network. Between numerous aggregation layer switches, it is
typically used to transfer traffic between servers, and
• As additional devices are added, energy
stateful network devices such server load balancers and
requirements and consumption will increase.
firewalls are attached to this layer. The switches that make
Currently, in traditional data centers, the top of the
up the access layer are typically used by servers to connect
rack (TOR) switch, also known as the fabric
to and communicate with one another. It contributes to
interconnect, is where all of the servers in a rack are
better network administration and is in charge of
connected. Each rack has a number of servers, either
exchanging any kind of traffic between servers, including
rack mount servers or chassis-based servers. On one
broadcast, multicast, and unicast.
side, this aggregation layer provides communication
between data centre clusters, allowing packet-based The bottom of the above illustration shows a server layer
East-West traffic to pass through, while on the other, stacked in server racks.
it also provides connectivity with the core network, Numerous Virtual machines are being operated by these
allowing North-South traffic to pass. As a result, the servers and assigned to various data centre applications. To
aggregation layer is the layer where problems arise manage and respond to requests from external clients
because 75% of the data center’s east-west traffic arriving through the internet, an application is typically
passes through it, causing a bandwidth bottleneck & linked to several public IP addresses. These requests are
an increase in latency. split up among the pool of servers for processing by
This article can be downloaded from here: www.ijaems.com 5
©2023 The Author(s). Published by Infogain Publication, This work is licensed under a Creative Commons Attribution 4.0 License.
http://creativecommons.org/licenses/by/4.0/
Hussain International Journal of Advanced Engineering, Management and Science, 9(9) -2023
partially address the shortcomings of the existing 3.3 OOO Switching Benefits
architecture while simultaneously being backwards • POD-based architecture is typically used in data
compatible with it. centres, which results in low use of computing
2.8.5 HFPFMAOS resources. However, optical switching enables
The data centre communication network architecture sharing of computing resources among several
(DCCN) called HFPFMAOS, is suggested as a solution to PODs for optimal effectiveness.
the various problems that data centres in the present face • Increases revenue by quickly deploying new
while maintaining the infrastructure. Particularly,in this services.
architecture, standard switches are used at the aggregation • Less power is lost compared to a traditional
layer. By monitoring all of the flows across the OF equipped electrical switch.
switches, this will give us consolidated network intelligence • On-demand capacity creation and reallocation.
and enable us to dynamically establish High bandwidth data • A remarkable increase in the data centre's
paths between the data centre servers whenever and operational efficiency as a result of the smooth
wherever they are needed. Additionally, it will lower functioning of applications and the efficient use of
network management overhead bytes and aid in quick computational resources and 3D optical MEMS
defect detection and correction. The conceptualized design switching technology.
is shown below. • Boost defect detection and quality of service.
3.4 Technology using OOO Switching
Different optical switching methods [17] are used by many
switches. The same are shown in the below table:
Table 3.1 Contrast of various Optical switching
technologies
packet filtering and topology change control. The OF 4.3.4 Flow Table
protocol is supported by nearly all of the well-known It is the fundamental component of logical switch design
vendors, including Cisco, HP, IBM, Brocade, and others. that determines the action to be taken on each incoming
There are two types of switches for OF or SDN compliance. packet by comparing it to a specific flow table made up of
4.3.1 Open Flow-Only Switches numerous flow entries. This packet might go via one or
These switches, which solely support OF operation and more flow tables that operate as pipelines. As demonstrated
process all packets using the OF pipeline, are also known as in fig. 4.5, a controller can add, remove, and alter flow
Pure SDN switches or SDN-only switches. entries from an OF switch's flow tables either proactively or
reactively (in reaction to an incoming packet).
4.3.2 Hybrid Switches with Open Flow Fig 4.5 Flow Table Entries Model
Fig 4.4 Architecture of Basic Open Flow (OF) Figure below depicts the packet processing flow chart.
4.3.5.1 Identifier
32-bit unique identifier identifies the group entry.
4.3.5.2 Type of Group
It is used to manage/maintain group types depicting them
Fig 4.7 Open Flow(OF)-Packet Processing Flow Chart. as “Required” and “Optional”.
4.3.5.3 Packet Counters
Table 4.1 Main components of Flow Entry Provide actual packets count which are generated and
processed by a group.
4.3.5.4 Action buckets
AB consists of complete flow of processes which are
4.3.4.1 Match Felds Values performed in a specific pattern for changing packets and
then sending them to a port. This collection of actions is
It is important here that not to pick any packets without always carried out as a set of acts. A group entry may have
matching field values as shown in the below table: zero or more buckets, with the exception of groups specified
Table 4.2 Match Field of Flow Entry as "Required: Indirect," which only include one bucket. If a
group doesn't have a bucket, it will immediately throw the
packets into the air.
The metadata of OpenFlow OF switches where BDDP Table 5.1 Topology Database
packets are received are included in "OFPT PACKET IN"
messages. Between OpenFlow OF switches and their
controller, messages are only transmitted in one direction.
These massages are also exchanged in the other direction.
Based on data from BDDP and metadata, the controller can
identify indirect multi-hop links between OpenFlow OF
switches after receiving these "OFPT PACKET IN" signals
and store them in its database to create a network topology.
Most crucially, the controller counts the number of hops
between OpenFlow OF switches using the TTL value. The
default time interval for this topology finding operation is 5
seconds. The total number of "OFPT PACKET IN"
messages received by the controller during this entire 5.1.3.3 Discovery of Host
discovery process is equal to double the number of "L" Two methods have been applied for available host
active links accessible in the domain, and this number may discovery which is connected to the OpenFlow OF
be calculated as below: switches.
As all active ports on all switches send BDDP "Packet Out"
However, the total messages “OFPT_PACKET_OUT” sent messages, I Ports for which OpenFlow OF Switches do not
by controller can be computed as: send "Packet In" messages are recognized as host ports
during link discovery. (ii) GARP message is sent by the
Host when Host is connected to an OpenFlow OF Switch.
HSRP and VRRP[24] use GARP to update the MAC
Number of Open Flow OF switches is denoted by S,
address table of L2 switches, In a broadcast domain GARP
whereas active ports are denoted by P. All switches has P. 4
is an advance notification mechanism to keep the controller
OpenFlow OF switches with 2 active ports each make up
aware about host discovery and inserting flow entries of
our reference topology, hence "OFPT PACKET OUT"
MAC address in the flow tables of Open Flow OF switches,
messages issued by the controller are 2+2+2+2=8 in total.
thus updating other hosts’ ARP tables before their ARP
The total number of BDDP OFPT PACKET OUT messages
requests are made, and finally updating ARP tables of other
sent by a controller to a switch can be reduced by
hosts when a new host is also connected to a
implementing OFDPv2[23], in which the Port ID TLV field
switch,however a host IP address or MAC address is
is set to 0 and will be ignored while the source MAC address
changed due to failover. Generated ARP requests are
field has been set with the MAC address of the port through
actually special ARP packets with the source that is (host)
which it is to be sent out. One "OFPT PACKET OUT"
IP and destination.The destination broadcast MAC address
message is sent by the controller for each switch, and the
(ff:ff:ff:ff:ff:ff) is present in the MAC address field and the
total number of messages transmitted is equal to the number
Ethertype field fixt to 0x0806. The parameters listed below
of switches determined by OFDPv2.
are part of a GARP request message:
• FF:FF:FF:FF:FF:FF:FF:FF is the destination MAC
As each OpenFlow OF switch's port is physically connected address (broadcast)
to a MEMS switch's port, we don't need link discovery to • Source MAC address: MAC address of the host
establish connectivity with the MAOS plane; instead, the Example of GARP is shown in the below figure. IP address
controller can create paths between any two points of a Source = IP address of a Destination: Host transmitting
dynamically, and topology database can statically store all GARP Type ,IP address is: ARP (0x0806).
the data related to flows and port connections.
Each host will ping its neighbor after GARP is finished to For providing maximum flexibility and support to diverse
verify connectivity and reachability. Currently, all hosts are types of data traffic passing through DCCN while supplying
aware of one another's MAC addresses and assigned IP delay-sensitive traffic with minimal latency, I will employ
addresses, which are visible in their ARP tables. To stop a hybrid model of flow table entries in this thesis that
BDDP packet propagation, SDN controllers disable link combines proactive and reactive modalities. The use of a
discovery on the OpenFlow OF switch ports to which hosts proactive model for flow entry is required for delay-
are connected after discovery. Similar to BPDU guard sensitive applications that demand low latency, for example
security function in conventional switches, this suppression while making audio/video calls, live radio/TV
also prevents BPDU propagation [27]. transmissions, financial banks transactions, and routine
Table 5.2 Host-1 linked with OSW1-1 - ARP heavy traffic like web browsing, data file/folder transfer,
and peer-to-peer traffic.
to add another new flow entry to its flow table through the
"Flow Mod" message. When a similar packet arrives at the
switch in the future and matches its fields and masks, this
flow entry guides the switch what to do now.In this way the
message informs the related switch to route any TCP
requests from Host1's IP address or MAC address to Host4's
IP address or MAC address and finally to Port 8.
Additionally, it tells the switch to release the packet it had
been buffering with the BufferID of 250 and to carry out the
Fig 5.10 Formula to calculate Link Utilization
instructions in this message. H4 replies by sending a
SYN/ACK packet to the switch, which receives it but finds
For each of the outgoing interfaces, the controller maintains no flow entry from Host2 to Host1 (yet another table miss).
a specific link utilisation threshold. When the threshold is In order to send this SYN/ACK packet to the controller for
exceeded, it consults its flowtable to filter flows with additional analysis, the switch encapsulates it in a "Packet
outgoing interfaces that have exceeded the threshold. In" message and gives it the reference buffer ID
Following the specification of ports, the SDN controller BufferID=251. In response, the controller instructs the
determines whether any ports are available before creating switch to add a flow entry to the flow table and perform
a Highband dat pathpath across the MEMS Plane. some action, which is to forward the SYN/ACK message to
port7. The controller also sends the switch a Packet Out and
If there is no match, the entry in the table is considered to
a Flow mod message.The remainder of the communication
be incorrect, and the controller is notified by default. The
between Host1 and Host2 would not reach the controller
controller instructs the switch to take certain actions. The
after all of this because switches have flow entries in their
"Flow Mod" message contains a variety of information,
flow tables that tell them what to do with packets. The
including Buffer IDs, Timeouts, Actions, Priorities, and
switch routed HTTP reply and ACK messages directly, as
more. Additionally, flow entries can be permanent or for a
demonstrated in the following figs:
limited period of time, and there are two types: "Hard
Timeout" and "Idle Timeout." Idle timeout refers to the
removal of an entry from the flow table if there is no
matching flow entry request during that time period. Hard
timeout refers to the maximum amount of time an entry can
remain in the flow table, regardless of whether a matching
entry is present. If Hard timeout is set to 0, it is deactivated.
As an illustration, in our reference topology diagram, Host
1 sends an HTTP request to Host 2 (let's say a web server)
following host discovery by GARP. It begins with a SYN
message. Host 1 sends a SYN message to the OSW1-1
switch, which checks its flow table upon receiving the
packet because it is the initial packet and likely has no flow
entries that match the packet. Table miss flow entry is the
term for this. Therefore, by encasing this TCP packet in a
"Packet IN" message, the switch passes it to the controller
by default. This Packet IN message contains the entire TCP
packet or its Buffer ID (for example, Buffer ID = 250, which
designates the location where the switch stores the whole
TCP message). Therefore, the controller will take a few
actions, such as returning a "Flow Mod" or "Packet Out"
message to the switch, where "Packet Out" includes
information regarding the switch's handling of that Fig 5.11 (a) HTTP request with Open Flow messaging
particular as well as the whole encapsulated TCP packet or
the reference buffer ID that the switch uses to store this
packet. Send the TCP SYNpacket reference with buffer
ID=250 out of port 8 to host 2 if the switch OSW1-1
receives a "Packet Out" message.Then the switch is directed
This article can be downloaded from here: www.ijaems.com 18
©2023 The Author(s). Published by Infogain Publication, This work is licensed under a Creative Commons Attribution 4.0 License.
http://creativecommons.org/licenses/by/4.0/
Hussain International Journal of Advanced Engineering, Management and Science, 9(9) -2023
Resultantly, the optical control plane gives messages to the is made up of two aggregation switches (SW3 & SW4), two
MAOS control plane, which is made up of MEMS-based access (TOR) switches (SW5 & SW6), and eight hosts
optical switches. (Hosts 1 through Host8) that produce traffic. POD2 consists
of 8 hosts, Host1 to Host8, 2 TOR switches, named SW9
and SW10, 2 aggregation switches, named as SW7 and
SW8, and 2 hosts.
6.2 Software used
The tools incorporated are shown in in the figure below:
6.3 Preparation/Implementation
Fig 6.8 OF switches based Network topology - Access as
Below actions must be taken to prepare for this setup: TOR switches
• Before running my Mininet and Open Day Light
virtual machines, I first downloaded and installed
Links between hosts and OF switches are 4 Mb/s with a 10
Oracle VM VirtualBox.
msec delay, while links between all traditional switches and
• Next, construct a VM in VirtualBox, download the between OF switches are 10 Mb/s with a 5 msec delay, as
Mininet virtual machine image, mount it on top of demonstrated in the following figures:
the fresh virtual machine, and then install Mininet. Host, Nodes, Links and interfaces verification can be done
• Third, create a second VirtualBox virtual machine by these commands, mininet>nodes, mininet>net and
and install the ODL controller setup. mininet>dump.
Mininet and its GUI program MiniEdit (running on o1 VM
on VirtualBox) is used by researcher, to plan out and
simulate the reference topology. I require X forwarding in
order to run Miniedit and connect to Mininet over SSH. To
do this, I've used Putty and XMing. Open DayLight, an
external SDN-based OpenFlow controller, was utilised by
me to operate OpenFlow virtual switches and MEMS
switches (ODL Beryllium). I have generated data flow from
hosts and servers and measured several performance
characteristics using the IPERF, such as link bandwidth and
network latency. I used Wireshark to investigate data Fig 6.9 Creating Links Switches
packets and analyse variety of protocol messages on several
interfaces during the whole network topology.
6.4 Applications
The real deployment of HFPFMAOS was gained in two
main steps