The Science DMZ: A Network Design Pattern For Data-Intensive Science
The Science DMZ: A Network Design Pattern For Data-Intensive Science
DOI 10.3233/SPR-140382
IOS Press
E-mail: [email protected]
Abstract. The ever-increasing scale of scientific data has become a significant challenge for researchers that rely on networks
to interact with remote computing systems and transfer results to collaborators worldwide. Despite the availability of high-
capacity connections, scientists struggle with inadequate cyberinfrastructure that cripples data transfer performance, and impedes
scientific progress. The Science DMZ paradigm comprises a proven set of network design patterns that collectively address these
problems for scientists. We explain the Science DMZ model, including network architecture, system configuration, cybersecurity,
and performance tools, that creates an optimized network environment for science. We describe use cases from universities,
supercomputing centers and research laboratories, highlighting the effectiveness of the Science DMZ model in diverse operational
settings. In all, the Science DMZ model is a solid platform that supports any science workflow, and flexibly accommodates
emerging network technologies. As a result, the Science DMZ vastly improves collaboration, accelerating scientific discovery.
Keywords: High performance networking, perfsonar, data-intensive science, network architecture, measurement
1058-9244/14/$27.50 © 2014 – IOS Press and the authors. All rights reserved
174 E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science
• Appropriate usage policies so that high- wide area network and the local area networks. The
performance applications are not hampered by wide area networks (or WANs) are often already op-
unnecessary constraints; timized and can accommodate large data flows up to
• An effective “on-ramp” for local resources to ac- 100 Gbps. However, the local area networks are still a
cess wide area network services; and choke point for these large data flows.
• Mechanisms for testing and measuring, thereby Local area networks are usually general-purpose
ensuring consistent performance. networks that support multiple missions, the first of
This paper will discuss the Science DMZ from its which is to support the organization’s business op-
development to its role in future technologies. First, erations including email, procurement systems, web
Section 2 will discuss the Science DMZ’s original browsing, and so forth. Second, these general networks
development in addressing the performance of TCP- must also be built with security that protects finan-
based applications. Second, Section 3 enumerates the cial and personnel data. Meanwhile, these networks
components of the Science DMZ model and how each are also used for research as scientists depend on this
component adds to the overall paradigm. Next, Sec- infrastructure to share, store, and analyze data from
tions 4 and 5 offer some sample illustrations of net- many different sources. As scientists attempt to run
works that vary in size and purpose. Following, Sec- their applications over these general-purpose networks,
tion 6 will discuss some examples of Science DMZ im- the result is often poor performance, and with the in-
plementations from the R&E community. And lastly, crease of data set complexity and size, scientists often
Section 7 highlights some future technological ad- wait hours, days, or weeks for their data to arrive.
vancements that will enhance the applicability of the Since many aspects of general-purpose networks are
Science DMZ design. difficult or impossible to change in the ways necessary
to improve their performance, the network architecture
must be adapted to accommodate the needs of science
2. Motivation applications without affecting mission critical business
and security operations. Some of these aspects that are
When developing the Science DMZ, several key difficult to change might include the size of the mem-
principles provided the foundation to its design. First, ory buffers for individual interfaces; mixed traffic pat-
these design patterns are optimized for science. This terns between mail and web traffic that would include
means the components of the system – including all science data; and emphasis on availability vs. perfor-
the equipment, software and associated services – are mance and what can be counted on over time for net-
configured specifically to support data-intensive sci- work availability.
ence. Second, the model is designed to be scalable in The Science DMZ model has already been im-
its ability to serve institutions ranging from large ex-
plemented at various institutions to upgrade these
perimental facilities to supercomputing sites to multi-
general-purpose, institutional networks. The National
disciplinary research universities to individual research
Science Foundation (NSF) recognized the Science
groups or scientists. The model also scales to serve a
DMZ as a proven operational best practice for univer-
growing number of users at those facilities with an in-
sity campuses supporting data-intensive science and
creasing and varying amount of data over time. Lastly,
the Science DMZ model was created with future inno- specifically identified this model as eligible for fund-
vation in mind by providing the flexibility to incorpo- ing through the Campus Cyberinfrastructure–Network
rate emerging network services. For instance, advances Infrastructure and Engineering Program (CC–NIE).2
in virtual circuit services, 100 Gigabit Ethernet, and This program was created in 2012 and has since been
the emergence of software-defined networking present responsible for implementing approximately 20 Sci-
new and exciting opportunities to improve scientific ence DMZs at different locations – thereby serving the
productivity. In this section, we will mostly discuss the needs of the science community. Another NSF solicita-
first principle since it is the driving mission for the Sci- tion was released in 2013 and awards to fund a similar
ence DMZ model. number of new Science DMZ’s are expected.
The first principle of the model is to optimize the
network for science. To do this, there are two entities 2 NSF’s CC–NIE Program: http://www.nsf.gov/pubs/2013/
or areas of the network that should be considered: the nsf13530/nsf13530.html.
E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science 175
Fig. 1. Graph shows the TCP throughput vs. round-trip time (latency) with packet loss between 10 Gbps connected hosts, as predicted by the
Mathis Equation. The topmost line (shown in purple) shows the throughput for TCP in a loss-free environment. (The colors are visible in the
online version of the article; http://dx.doi.org/10.3233/SPR-140382.)
2.1. TCP performance second due to the loss rate. While this only resulted
in an overall drop of throughput of 450 Kbps (on the
The Transmission Control Protocol (TCP) [15] of device itself), it reduced the end-to-end TCP perfor-
the TCP/IP protocol suite is the primary transport pro- mance far more dramatically as demonstrated in Fig. 1.
tocol used for the reliable transfer of data between ap- This packet loss was not being reported by the router’s
plications. TCP is used for email, web browsing, and internal error monitoring, and was only noticed using
similar applications. Most science applications are also the owamp active packet loss monitoring tool, which is
built on TCP, so it is important that the networks are part of the perfSONAR Toolkit.4
able to work with these applications (and TCP) to op-
Because TCP interprets the loss as network conges-
timize the network for science.
tion, it reacts by rapidly reducing the overall sending
TCP is robust in many respects – in particular it has
rate. The sending rate then slowly recovers due to the
sophisticated capabilities for providing reliable data
delivery in the face of packet loss, network outages, dynamic behavior of the control algorithms. Network
and network congestion. However, the very mecha- performance can be negatively impacted at any point
nisms that make TCP so reliable also make it per- during the data transfer due to changing conditions in
form poorly when network conditions are not ideal. In the network. This problem is exacerbated as the la-
particular, TCP interprets packet loss as network con- tency increases between communicating hosts. This is
gestion, and reduces its sending rate when loss is de- often the case when research collaborations sharing
tected. In practice, even a tiny amount of packet loss is data are geographically distributed. In addition, feed-
enough to dramatically reduce TCP performance, and back regarding the degraded performance takes longer
thus increase the overall data transfer time. When ap- to propagate between the communicating hosts.
plied to large tasks, this can mean the difference be- The relationship between latency, data loss, and net-
tween a scientist completing a transfer in days rather work capability was described by Mathis et al. as a
than hours or minutes. Therefore, networks that sup- mechanism to predict overall throughput [12]. The
port data-intensive science must provide TCP-based “Mathis Equation” states that maximum TCP through-
applications with loss-free service if TCP-based appli- put is at most:
cations are to perform well in the general case.
As an example of TCP’s sensitivity, consider the fol- maximum segment size 1
lowing case. In 2012, Department of Energy’s (DOE) ×√ .
round-trip time packet loss rate
Energy Sciences Network (ESnet) had a failing 10
Gbps router line card that was dropping 1 out of 22,000 (1)
packets, or 0.0046% of all traffic. Assuming the line
card was working at peak efficiency, or 812,744 regu- Figure 1 shows the theoretical rate predicted by the
lar sized frames per second,3 37 packets were lost each Mathis Equation, along with the measured rate for both
TCP-Reno and TCP-Hamilton across ESnet. These
3 Performance Metrics, http://www.cisco.com/web/about/security/
intelligence/network_performance_metrics.html. 4 perfSONAR Toolkit: http://psps.perfsonar.net.
176 E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science
tests are between 10 Gbps connected hosts configured There are several reasons to separate the high-
to use 9 KByte (“Jumbo Frame”) Maximum Transmis- performance science traffic from the rest of the net-
sion Units (MTUs). work. The support of high-performance applications
This example is indicative of the current opera- can involve the deployment of highly capable equip-
tional reality in science networks. TCP is used for ment that would be too expensive to use throughout the
the vast majority of high-performance science applica- general-purpose network but that has necessary fea-
tions. Since TCP is so sensitive to loss, a science net- tures such as high-performance filtering capabilities,
work must provide TCP with a loss-free environment, sufficient buffering for burst capacity, and the ability to
end-to-end. This requirement, in turn, drives a set of accurately account for packets that traverse the device.
design decisions that are key components of the Sci- In some cases, the configuration of the network devices
ence DMZ model. must be changed to support high-speed data flows – an
example might be conflict between quality of service
settings for the support of enterprise telephony and the
3. The science DMZ design pattern burst capacity necessary to support long-distance high-
performance data flows. In addition, the location pat-
tern makes the application of the appropriate security
The overall design pattern or paradigm of the Sci-
pattern significantly easier (see Section 3.4).
ence DMZ is comprised of four sub-patterns. Each of
The location design pattern can also significantly
these sub-patterns offers repeatable solutions for four
reduce the complexity of the portion of the network
different areas of concern: proper location (in network
used for science applications. Troubleshooting is time-
terms) of devices and connections; dedicated systems; consuming, and there is a large difference in opera-
performance measurement; and appropriate security tional cost and time-to-resolution between verifying
policies. These four sub-patterns will be discussed in the correct operation of a small number of routers and
the following subsections. switches and tracing the science flow through a large
number of network devices in the general-purpose net-
3.1. Proper location to reduce complexity work of a college campus. For this reason, the Sci-
ence DMZ is typically located as close to the network
The physical location of the Science DMZ (or “lo- perimeter as possible, i.e. close to or directly connected
cation design pattern") is important to consider during to the border router that connects the research institu-
the deployment process. The Science DMZ is typically tion’s network to the wide area science network.
deployed at or near the network perimeter of the in-
stitution. The reason for this is that it is important to 3.2. Dedicated systems: The Data Transfer Node
involve as few network devices as reasonably possible (DTN)
in the data path between the experiment at a science
facility, the Science DMZ, and the WAN. Systems used for wide area science data transfers
Network communication between applications run- perform far better if they are purpose-built for and ded-
ning on two hosts traverses, by definition, the hosts icated to this function. These systems, which we call
themselves and the entire network infrastructure be- data transfer nodes (DTNs), are typically PC-based
tween the hosts. Given the sensitivity of TCP to packet Linux servers constructed with high quality compo-
loss (as discussed in Section 2.1), it is important to en- nents and configured specifically for wide area data
sure that all the components of the network path be- transfer. The DTN also has access to storage resources,
tween the hosts are functioning properly and config- whether it is a local high-speed disk subsystem, a con-
nection to a local storage infrastructure, such as a stor-
ured correctly. Wide area science networks are typi-
age area network (SAN), or the direct mount of a high-
cally engineered to perform well for science applica-
speed parallel file system such as Lustre5 or GPFS.6
tions, and in fact the Science DMZ model assumes that
The DTN runs the software tools used for high-speed
the wide area network is doing its job. However, the
data transfer to remote systems. Some typical software
local network is often complex, and burdened with the
packages include GridFTP7 [3] and its service-oriented
compromises inherent in supporting multiple compet-
ing missions. The location design pattern accomplishes 5 Lustre, http//www.lustre.org/.
two things. The first is separation from the rest of the 6 GPFS, http://www.ibm.com/systems/software/gpfs/.
general network, and the second is reduced complexity. 7 GridFTP, http://www.globus.org/datagrid/gridftp.html.
E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science 177
front-end Globus Online8 [4], discipline-specific tools rity perspective, and this makes the appropriate secu-
such as XRootD,9 and versions of default toolsets such rity policy easier to apply (see Section 3.4).
as SSH/SCP with high-performance patches10 applied. Because the design and tuning of a DTN can be
DTNs are widely applicable in diverse science en- time-consuming for small research groups, ESnet has
vironments. For example, DTNs are deployed to sup- a DTN Tuning guide19 and a Reference DTN Imple-
port Beamline 8.3.2 at Berkeley Lab’s Advanced Light mentation guide.20 The typical engineering trade-offs
Source,11 and as a means of transferring data to and between cost, redundancy, performance and so on. ap-
from a departmental cluster. On a larger scale, sets ply when deciding on what hardware to use for a DTN.
of DTNs are deployed at supercomputer centers (for In general, it is recommended that DTNs be procured
example at the DOE’s Argonne Leadership Comput- and deployed such that they can be expanded to meet
ing Facility,12 the National Energy Research Scientific future storage requirements.
Computing Center,13 and Oak Ridge Leadership Com-
puting Facility14 ) to facilitate high-performance trans- 3.3. Performance monitoring
fer of data both within the centers and to remote sites.
At even larger scales, large clusters of DTNs provide Performance monitoring is critical to the discov-
data service to the Large Hadron Collider (LHC)15 col- ery and elimination of so-called “soft failures” in the
laborations. The Tier-116 centers deploy large numbers network. Soft failures are problems that do not cause
of DTNs to support thousands of scientists. These are a complete failure that prevents data from flowing
systems dedicated to the task of data transfers so that (like a fiber cut), but causes poor performance. Ex-
they provide reliable, high-performance service to sci- amples of soft failures include packet loss due to fail-
ence applications.17 ing components; dirty fiber optics; routers forward-
DTNs typically have high-speed network interfaces, ing packets using the management CPU rather than
but the key is to match the DTN to the capabilities of the high-performance forwarding hardware; and inad-
the wide area network infrastructure. For example, if equate hardware configuration. Soft failures often go
the network connection from the site to the WAN is undetected for many months or longer, since most net-
1 Gigabit Ethernet, a 10 Gigabit Ethernet interface on work management and error reporting systems are op-
the DTN may be counterproductive. The reason for this timized for reporting “hard failures”, such as loss of
is that a high-performance DTN can overwhelm the a link or device. Also, many scientists do not know
slower wide area link causing packet loss. what level of performance to expect, and so they do not
The set of applications that run on a DTN is typi- know when to alert knowledgeable staff about a poten-
cally limited to parallel data transfer applications like tial problem.
GridFTP or FDT.18 In particular, user-agent applica- A perfSONAR host [16] helps with fault diagno-
tions associated with general-purpose computing and sis on the Science DMZ. It offers end-to-end testing
business productivity (e.g., email clients, document ed- with collaborating sites that have perfSONAR tools in-
itors, media players) are not installed. This is for two stalled, which allows for multi-domain troubleshoot-
reasons. First, the dedication of the DTN to data trans- ing. perfSONAR is a network monitoring software
fer applications produces more consistent behavior and suite designed to conduct both active and passive net-
avoids engineering trade-offs that might be part of sup- work measurements, convert these to a standard for-
porting a larger application set. Second, data transfer mat, and then publish the data so it is publicly acces-
applications are relatively simple from a network secu- sible. The perfSONAR host can run continuous checks
8 Globus Online, https://www.globusonline.org/.
for latency changes and packet loss using OWAMP,21
9 XRootD, http://xrootd.slac.stanford.edu/. as well as periodic “throughput” tests (a measure of
10 HPN-SSH, http://www.psc.edu/networking/projects/hpn-ssh/. available network bandwidth) using BWCTL.22 If a
11 LBNL ALS, http://www-als.lbl.gov. problem arises that requires a network engineer to
12 ALCF, https://www.alcf.anl.gov. troubleshoot the routing and switching infrastructure,
13 NERSC, http://www.nersc.gov.
14 OLCF, http://www.olcf.ornl.gov/. 19 DTN Tuning, http://fasterdata.es.net/science-dmz/DTN/
15 LHC, http://lhc.web.cern.ch/lhc/. tuning/.
16 US/LHC, http://www.uslhc.us/The_US_and_the_LHC/ 20 Reference DTN, http://fasterdata.es.net/science-dmz/data-
Computing. transfer-node-reference-implementation/.
17 LHCOPN, http://lhcopn.web.cern.ch/lhcopn/. 21 OWAMP, http://www.internet2.edu/performance/owamp/.
18 FTD, http://monalisa.cern.ch/FDT/. 22 BWCTL, http://www.internet2.edu/performance/bwctl/.
178 E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science
the application’s traffic. The firewall administrator asks 6.1. University of Colorado, Boulder
for the IP addresses of the communicating hosts, and
the TCP ports that will be used by the hosts to commu- The University of Colorado, Boulder campus was an
nicate. Armed with that information, the firewall ad- early adopter of Science DMZ technologies. Their core
ministrator configures the firewall to permit the traf- network features an immediate split into a protected
fic. Filtering based on IP address and TCP port num- campus infrastructure (beyond a firewall), as well as
ber can be done on the Science DMZ switch or router a research network (RCNet) that delivers unprotected
with ACLs. When done with ACLs on a modern switch functionality directly to campus consumers. Figure 6
or router, the traffic does not need to traverse a fire- shows the basic breakdown of this network, along with
wall at all. This is a key point: by running a limited set the placement of measurement tools provided by perf-
of applications on the Science DMZ DTNs, the appli-
SONAR.
cation profile is such that the Science DMZ can typ-
The physics department, a participant in the Com-
ically be defended well without incurring the perfor-
pact Muon Solenoid (CMS)24 experiment affiliated
mance penalties of a firewall. This is especially true if
with the LHC project, is a heavy user of campus net-
the ACLs are used in combination with intrusion de-
work resources. It is common to have multiple streams
tection systems or other advanced security tools. How-
of traffic approaching an aggregate of 5 Gbps affiliated
ever, an intrusion detection system should be used even
if a firewall is present. with this research group. As demand for resources in-
Aggregation (“fan-in”) problems are related to the creased, the physics group connected additional com-
firewall problem in that they too result from the com- putation and storage to their local network. Figure 7
bination of the burstiness of TCP traffic and small shows these additional 1 Gbps connections as they en-
buffers on network devices. However, the fan-in prob- tered into the portion of the RCNet on campus.
lem arises when multiple traffic flows entering a switch Despite the initial care in the design of the network,
or router from different ingress interfaces are destined overall performance began to suffer during heavy use
for a common egress interface. If the speed of the sum times on the campus. Passive and active perfSONAR
of the bursts arriving at the switch is greater than the monitoring alerted that there was low throughput to
speed of the device’s egress interface, the device must
buffer the extra traffic or drop it. If the device does
not have sufficient buffer space, it must drop some of
the traffic, causing TCP performance problems. This
situation is particularly common in inexpensive LAN
switches. Since high-speed packet memory is expen-
sive, cheap switches often do not have enough buffer
space to handle anything except LAN traffic. Note that Fig. 6. University of Colorado campus network, showing RCNet
the fan-in problem is not unique to coincident bursts. If connected at the perimeter as a Science DMZ. (Colors are visi-
a burst from a single flow arrives at a rate greater than ble in the online version of the article; http://dx.doi.org/10.3233/
the rate available on the egress interface due to existing SPR-140382.)
non-bursty traffic flows, the same problem exists.
6. Use cases
torical weather forecasts with the same current version DMZ architecture at NERSC. Most recently, it has
of NOAA’s Global Ensemble Forecast System (GEFS). enabled high-speed multi-terabyte transfers between
Among the advantages associated with a long refore- SLAC Linear Accelerator National Lab’s Linac Co-
cast data set, model forecast errors can be diagnosed herent Light Source and NERSC to support protein
from the past forecasts and corrected, thereby dramat- crystallography experiments as well as transfers be-
ically increasing the forecast skill, especially in fore- tween Beamline 8.3.2 at Berkeley Lab’s Advanced
casts of relatively rare events and longer-lead forecast. Light Source and NERSC in support of X-ray tomog-
In 2010, the NOAA team received an allocation raphy experiments.
of 14.5 million processor hours at NERSC to per-
form this work. In all, the 1984–2012 historical GEFS
dataset totaled over 800 TB, stored on the NERSC 7. Future technologies
HPSS archival system. Of the 800 TB at NERSC, the
NOAA team sought to bring about 170 TB back to
In addition to solving today’s network performance
NOAA Boulder for further processing and to make it
problems, the Science DMZ model also makes it easier
more readily available to other researchers. When the
to experiment and integrate with tomorrow’s technolo-
NOAA team tried to use an FTP server located behind
gies. Technologies such as dynamic “virtual circuits”,
NOAA’s firewall for the transfers, they discovered that
software-defined networking (SDN), and 40/100 Gbps
data trickled in at about 1–2 MB/s.
ethernet can be deployed in the Science DMZ, elimi-
Working with ESnet and NERSC, the NOAA team
nating the need to deploy these technologies deep in-
leveraged the Science DMZ design pattern to set up a
side campus infrastructure.
new dedicated transfer node enabled with Globus On-
line to create a data path unencumbered by legacy fire-
walls. Immediately the team saw a throughput increase 7.1. Virtual circuits
of nearly 200 times. The team was able to transfer 273
files with a total size of 239.5 GB to the NOAA DTN Virtual circuit services, such as the ESnet-developed
in just over 10 minutes – approximately 395 MB/s. On-demand Secure Circuits and Reservation System,
or OSCARS platform [7,14], can be used to connect
6.4. National Energy Research Scientific Computing wide area layer-2 circuits directly to DTNs, allow-
Center ing the DTNs to receive the benefits of the bandwidth
reservation, quality of service guarantees, and traf-
In 2009, both NERSC and OLCF installed DTNs to fic engineering capabilities. The campus or lab “inter-
enable researchers who use their computing resources domain” controller (IDC)25 can provision the local
to move large data sets between each facility’s mass switch and initiate multi-domain wide area virtual cir-
storage systems. As a result, WAN transfers between cuit connectivity to provide guaranteed bandwidth be-
NERSC and OLCF increased by at least a factor of tween DTN’s at multiple institutions. An example of
20 for many collaborations. As an example, a com- this configuration is the NSF-funded Development of
putational scientist in the OLCF Scientific Comput- Dynamic Network System (DYNES) [17] project that
ing Group who was researching the fundamental nu- is supporting a deployment of approximately 60 uni-
clear properties of carbon-14, in collaboration with versity campuses and regional networks across the US.
scientists from Lawrence Livermore National Labora- Virtual circuits also enable the use of new data transfer
tory (LLNL) and Iowa State University, had previously protocols such as RDMA (remote direct memory ac-
waited more than an entire workday for a single 33 GB cess) over Converge Ethernet (RoCE) [5] on the Sci-
input file to transfer – just one of the 20 files of sim- ence DMZ DTNs. RoCE has been demonstrated to
ilar size that needed to be moved between the sites. work well over a wide area network, but only on a guar-
With the improved infrastructure, those researchers anteed bandwidth virtual circuit with minimal com-
were immediately able to improve their transfer rate peting traffic [11]. Kissel et al. show that RoCE can
to 200 MB/s enabling them to move all 40 TB of data achieve the same performance as TCP (39.5 Gbps for
between NERSC and OLCF in less than three days. a single flow on a 40GE host), but with 50 times less
Since 2009, several science collaborations includ- CPU utilization.
ing those in astrophysics, climate, photon science, ge-
nomics and others have benefitted from the Science 25 IDC, http://www.controlplane.net/.
184 E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science
100 Gigabit Ethernet (GE) technology is being de- The Science DMZ model has its roots in opera-
ployed by research networks around the world, to sup- tional practices developed over years of experience,
port data-intensive science. The NSF CC–NIE pro- and incorporates aspects of network architecture, net-
gram is increasing the rate of 100GE deployment at US work security, performance tuning, system design, and
campuses with solicitations offered in 2012 and 2013. application selection. The Science DMZ, as a design
While 100GE promises the ability to support next- pattern, has already been successfully deployed at mul-
generation instruments and facilities, and to conduct tiple sites across the US, and many through NSF fund-
scientific analysis of distributed data sets at unprece- ing. The Science DMZ model and its contributing tech-
dented scale, 100GE technology poses significant chal- nologies are well-tested and have been effectively used
lenges for the general-purpose networks at research at supercomputer centers, national laboratories, and
institutions. Once a site is connected to a 100GE back- universities as well as in large-scale scientific experi-
bone, it would be very costly to distribute this new in- ments.
creased bandwidth across internal campus infrastruc- The Science DMZ model provides a conceptual
ture. With the Science DMZ model, all hosts need- framework for the deployment of networks and
ing the increased bandwidth are near the border router, network-enabled tools and systems for the effective
making it much easier to benefit from the 100GE con- support of data-intensive science. With many science
nection. collaborations moving to large-scale or distributed ex-
periments, the purpose of sharing best practices is be-
coming more important. This paper shares our work
7.3. Software-defined networking
in developing the Science DMZ for the larger science
community.
Testing and deploying software defined networking,
particularly the use of OpenFlow as a platform [13],
is a timely example of how the Science DMZ model Acknowledgements
could be used for exploring and hardening new tech-
nologies. The authors would like to thank NOAA, NERSC,
Software-defined networking concepts and produc- the Pennsylvania State University, and the University
tion uses of OpenFlow are still in their early stages of Colorado, Boulder, for their contributions to this
of adoption by the community. Many innovative ap- work.
proaches are still being investigated to develop best The authors wish to acknowledge the vision of the
practices for the deployment and integration of these National Science Foundation for its support of the CC–
services in production environments. ESnet and its col- NIE program.
laborators at Indiana University have demonstrated an This manuscript has been authored by an author at
OpenFlow-based Science DMZ architecture that inter- Lawrence Berkeley National Laboratory under Con-
operates with a virtual circuit service like OSCARS. It tract No. DE-AC02-05CH11231 with the U.S. Depart-
is easy to set up an OSCARS virtual circuit across the ment of Energy. The U.S. Government retains, and
WAN, but plumbing the circuit all the way to the end the publisher, by accepting the article for publica-
host must be done by hand. OpenFlow simplifies this tion, acknowledges, that the U.S. Government retains
process. a non-exclusive, paid-up, irrevocable, world-wide li-
Another promising use of OpenFlow is as a mech- cense to publish or reproduce the published form of
anism to dynamically modify the security policy for this manuscript, or allow others to do so, for U.S. Gov-
large flows between trusted sites. Multiple groups have ernment purposes.
demonstrated the use of OpenFlow to dynamically by-
pass the firewall (e.g., Kissel et al.’s research on SDN
with XSP [10]). Further, one could also use OpenFlow Disclaimer
along with an intrusion detection system (IDS) to send
the connection setup traffic to the IDS for analysis, and This document was prepared as an account of work
then once the connection is verified allow the flow to sponsored by the United States Government. While
bypass the firewall and the IDS. this document is believed to contain correct informa-
E. Dart et al. / The Science DMZ: A network design pattern for data-intensive science 185
tion, neither the United States Government nor any [7] C. Guok, D. Robertson, M. Thompson, J. Lee, B. Tierney and
agency thereof, nor the Regents of the University of W. Johnston, Intra and interdomain circuit provisioning us-
California, nor any of their employees, makes any war- ing the OSCARS reservation system, in: Third International
Conference on Broadband Communications Networks and Sys-
ranty, express or implied, or assumes any legal respon- tems, IEEE/ICST, October 2006.
sibility for the accuracy, completeness, or usefulness [8] V. Jacobson, R. Braden and D. Borman, TCP Extensions for
of any information, apparatus, product, or process dis- High Performance, RFC 1323 (Proposed Standard), May 1992.
closed, or represents that its use would not infringe pri- [9] W.E. Johnston, E. Dart, M. Ernst and B. Tierney, Enabling high
vately owned rights. Reference herein to any specific throughput in widely distributed data management and analy-
commercial product, process, or service by its trade sis systems: Lessons from the LHC, in: TERENA Networking
name, trademark, manufacturer, or otherwise, does not Conference (TNC) 2013, June 2013.
necessarily constitute or imply its endorsement, rec- [10] E. Kissel, G. Fernandes, M. Jaffee, M. Swany and M. Zhang,
ommendation, or favoring by the United States Gov- Driving software defined networks with XSP, in: Workshop on
Software Defined Networks (SDN’12), International Confer-
ernment or any agency thereof, or the Regents of the ence on Communications (ICC), IEEE, June 2012.
University of California. The views and opinions of [11] E. Kissel, B. Tierney, M. Swany and E. Pouyoul, Efficient data
authors expressed herein do not necessarily state or transfer protocols for big data, in: Proceedings of the 8th Inter-
reflect those of the United States Government or any national Conference on eScience, IEEE, July 2012.
agency thereof or the Regents of the University of Cal- [12] M. Mathis, J. Semke, J. Mahdavi and T. Ott, The macro-
ifornia. scopic behavior of the tcp congestion avoidance algorithm,
SIGCOMM Comput. Commun. Rev. 27(3) (1997), 67–82.
[13] N. McKeown, T. Anderson, H. Balakrishnan, G. Parulkar,
References L. Peterson, J. Rexford, S. Shenker and J. Turner, Openflow:
enabling innovation in campus networks, SIGCOMM Comput.
[1] C. Alexander, The Timeless Way of Building, Oxford Univ. Commun. Rev. 38(2) (2008), 69–74.
Press, New York, 1979. [14] I. Monga, C. Guok, W.E. Johnston and B. Tierney, Hybrid net-
[2] C. Alexander, S. Ishikawa and M. Silverstein, A Pattern Lan- works: Lessons learned and future challenges based on ESnet4
guage: Towns, Buildings, Construction, Oxford Univ. Press, experience, IEEE Communications Magazine, May 2011.
New York, 1977. [15] J. Postel, Transmission Control Protocol, Request for Com-
[3] W. Allcock, J. Bresnahan, R. Kettimuthu, M. Link, C. Du- ments (Standard) 793, Internet Engineering Task Force,
mitrescu, I. Raicu and I. Foster, The globus striped GridFTP September 1981.
framework and server, in: Proceedings of the 2005 ACM/IEEE [16] B. Tierney, J. Boote, E. Boyd, A. Brown, M. Grigoriev, J. Met-
Conference on Supercomputing, SC’05, IEEE Computer Soci- zger, M. Swany, M. Zekauskas and J. Zurawski, perfSONAR:
ety, Washington, DC, USA, 2005, p. 54. Instantiating a global network measurement framework, in:
[4] B. Allen, J. Bresnahan, L. Childers, I. Foster, G. Kandaswamy, SOSP Workshop on Real Overlays and Distributed Systems
R. Kettimuthu, J. Kordas, M. Link, S. Martin, K. Pickett et al., (ROADS’09), Big Sky, MT, USA, ACM, October 2009.
Software as a service for data scientists, Communications of
[17] J. Zurawski, R. Ball, A. Barczyk, M. Binkley, J. Boote,
the ACM 55(2) (2012), 81–88.
E. Boyd, A. Brown, R. Brown, T. Lehman, S. McKee,
[5] I.T. Association, InfiniBand. Architecture Specification Re-
B. Meekhof, A. Mughal, H. Newman, S. Rozsa, P. Sheldon,
lease 1.2.1 Annex A16: RoCE, 2010.
A. Tackett, R. Voicu, S. Wolff and X. Yang, The dynes instru-
[6] E. Gamma, R. Helm, R. Johnson and J. Vlissides, Design
ment: A description and overview, Journal of Physics: Confer-
Patterns: Elements of Reusable Object-Oriented Software,
ence Series 396(4) (2012), 042065.
Addison-Wesley Longman Publishing, Boston, MA, USA,
1995.
Advances in Journal of
Industrial Engineering
Multimedia
Applied
Computational
Intelligence and Soft
Computing
The Scientific International Journal of
Distributed
Hindawi Publishing Corporation
World Journal
Hindawi Publishing Corporation
Sensor Networks
Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation
http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014
Advances in
Fuzzy
Systems
Modelling &
Simulation
in Engineering
Hindawi Publishing Corporation
Hindawi Publishing Corporation Volume 2014 http://www.hindawi.com Volume 2014
http://www.hindawi.com
International Journal of
Advances in Computer Games Advances in
Computer Engineering Technology Software Engineering
Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation Hindawi Publishing Corporation
http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014 http://www.hindawi.com Volume 2014
International Journal of
Reconfigurable
Computing