3.6. Defining the Cluster Configuration

Documentation

VoltDB Home » Documentation » Using VoltDB

3.6. Defining the Cluster Configuration

An important aspect of a VoltDB database is the physical layout of the cluster that runs the database. You define the cluster configuration in the deployment file. The deployment file is an XML file, which you specify when you start the database to establish the correct cluster topology. The basic syntax of the deployment file is as follows:

<?xml version="1.0"?>
<deployment>
   <cluster hostcount="n"
            kfactor="n"
   />
</deployment>

The attributes of the <cluster> tag define the physical layout of the hardware that will run the database. The key attributes are:

  • hostcount — specifies the number of nodes in the cluster.

  • kfactor — specifies the K-safety value to use for durability when creating the database. The K-safety value controls the duplication of database partitions. This attribute is optional, so if you do not specify a value the default is zero, which means there is no partition duplication. See Chapter 10, Availability for more information about K-safety.

In the simplest case — when running on a single node with no special options enabled — you can skip the deployment file altogether on the voltdb command line. If you do not specify a deployment file or host, VoltDB defaults to one node, eight execution sites per host, and a K-safety value of zero.

The deployment file is used to enable and configure many other runtime options related to the database, which are described later in this book. For example, the deployment file can specify:

For the complete deployment file features, see Appendix E, Deployment File (deployment.xml).

3.6.1. Determining How Many Sites per Host

There is very little penalty for allocating more sites than needed for the partitions the database will use (except for incremental memory usage). Consequently, VoltDB defaults to eight sites per node to provide reasonable performance on most modern system configurations. This default does not normally need to be changed. However, for systems with a large number of available processes (16 or more) or older machines with fewer than 8 processors and limited memory, you may wish to tune the sitesperhost attribute.

The number of sites needed per node is related to the number of processor cores each system has, the optimal number being approximately 3/4 of the number of CPUs reported by the operating system. For example, if you are using a cluster of dual quad-core processors (in other words, 8 cores per node), the optimal number of partitions is likely to be 6 or 7 sites per node.

<?xml version="1.0"?>
<deployment>
   <cluster . . .
            sitesperhost="6"
   />
</deployment>

For systems that support hyperthreading (where the number of physical cores support twice as many threads), the operating system reports twice the number of physical cores. In other words, a dual quad-core system would report 16 virtual CPUs. However, each partition is not quite as efficient as on non- hyperthreading systems. So the optimal number of sites is more likely to be between 10 and 12 per node in this situation.

Because there are no hard and set rules, the optimal number of sites per node is best calculated by actually benchmarking the application to see what combination of cores and sites produces the best results. However, it is important to remember that all nodes in the cluster will use the same number of sites. So the best performance is achieved by using a cluster with all nodes having the same physical architecture (i.e. cores).

3.6.2. Configuring Paths for Runtime Features

An important aspect of some runtime features is that they make use of disk resources for persistent storage across sessions. For example, automatic snapshots need a directory for storing snapshots of the database contents. Similarly, export uses disk storage for writing overflow data if the export connector cannot keep up with the export queue.

You can specify individual paths for each feature, or you can specify a root directory where VoltDB will create subfolders for each feature as needed. If you do not specify a root path or a specific feature path, the root path defaults to ./voltdbroot in the current default directory and VoltDB creates the directory (and subfolders) as needed.

To specify a common root, use the <voltdbroot> tag (as a child of <paths>) to specify where VoltDB will store disk files. If you specify a root directory path, the directory must exist and the process running VoltDB must have write access to it. VoltDB does not attempt to create an explicitly named root directory path if it does not exist. For example, the following <paths> tag specifies /opt/voltdb as the root directory:

<paths>
   <voltdbroot path="/opt/voltdb" />
</paths>

You can also identify specific path locations for individual features including:

  • <commandlog>

  • <commandlogsnapshot>

  • <exportoverflow>

  • <snapshots>

If you name a specific feature path and it does not exist, VoltDB will attempt to create it for you. For example, the <exportoverflow> path contains temporary data which can be deleted periodically. The following excerpt from a deployment file specifies /opt/voltdb as the default root but /opt/overflow as the directory for export overflow.

<paths>
   <voltdbroot path="/opt/voltdb" />
   <exportoverflow path="/opt/overflow" />
</paths>

3.6.3. Verifying your Hardware Configuration

The deployment file defines the expected configuration of your database cluster. However, there are several important aspects of the physical hardware and operating system configuration that you should be aware of before running VoltDB:

  • VoltDB can operate on heterogeneous clusters. However, best performance is achieved by running the cluster on similar hardware with the same type of processors, number of processors, and amount of memory on each node.

  • All nodes must be able to resolve the IP addresses and host names of the other nodes in the cluster. That means they must all have valid DNS entries or have the appropriate entries in their local hosts file.

  • You must run the Network Time Protocol (NTP) on all of the cluster nodes, preferably synchronizing against the same local time server. If the time skew between nodes in the cluster is greater than 100 milliseconds, VoltDB cannot start the database.

  • It is strongly recommended that you run NTP with the -x argument. Using ntpd -x stops the server from adjusting time backwards for all but very large increments. If the server time moves backward, VoltDB must pause and wait for time to catch up.