Pve Admin Guide - 2
Pve Admin Guide - 2
File /etc/apt/sources.list.d/ceph.list
This Ceph repository contains the Ceph packages before they are moved to the main repository. It is used
to test new Ceph releases on Proxmox VE.
File /etc/apt/sources.list.d/ceph.list
If Ceph is deployed this repository is needed for the upgrade from Proxmox VE 5.x to Proxmox VE 6.0. It
provides packages for the older Ceph Luminous release for Proxmox VE 6.0.
The Upgrade 5.x to 6.0 document explains how to use this repository in detail.
File /etc/apt/sources.list.d/ceph.list
3.1.9 SecureApt
The Release files in the repositories are signed with GnuPG. APT is using these signatures to verify that all
packages are from a trusted source.
If you install Proxmox VE from an official ISO image, the key for verification is already installed.
If you install Proxmox VE on top of Debian, download and install the key with the following commands:
# wget http://download.proxmox.com/debian/proxmox-ve-release-6.x.gpg -O / ←-
etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
# sha512sum /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
acca6f416917e8e11490a08a1e2842d500b3a5d9f322c6319db0927b2901c3eae23cfb5cd5df6facf2b
/etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
or:
# md5sum /etc/apt/trusted.gpg.d/proxmox-ve-release-6.x.gpg
f3f6c5a3a67baf38ad178e5ff1ee270c /etc/apt/trusted.gpg.d/proxmox-ve-release ←-
-6.x.gpg
Proxmox provides updates on a regular basis for all repositories. To install updates use the web-based GUI
or the following CLI commands:
# apt-get update
# apt-get dist-upgrade
Note
The APT package management system is very flexible and provides many features, see man apt-get,
or [Hertzog13] for additional information.
Tip
Regular updates are essential to get the latest patches and security related fixes. Major system upgrades
are announced in the Proxmox VE Community Forum.
Network configuration can be done either via the GUI, or by manually editing the file /etc/network/interfac
which contains the whole network configuration. The interfaces(5) manual page contains the com-
plete format description. All Proxmox VE tools try hard to keep direct user modifications, but using the GUI
is still preferable, because it protects you from errors.
Once the network is configured, you can use the Debian traditional tools ifup and ifdown commands to
bring interfaces up and down.
Proxmox VE Administration Guide 29 / 464
Proxmox VE does not write changes directly to /etc/network/interfaces. Instead, we write into a
temporary file called /etc/network/interfaces.new, this way you can do many related changes at
once. This also allows to ensure your changes are correct before applying, as a wrong network configuration
may render a node inaccessible.
With the default installed ifupdown network managing package you need to reboot to commit any pending
network changes. Most of the time, the basic Proxmox VE network setup is stable and does not change
often, so rebooting should not be required often.
With the optional ifupdown2 network managing package you also can reload the network configuration
live, without requiring a reboot.
Note
ifupdown2 cannot understand OpenVSwitch syntax, so reloading is not possible if OVS interfaces are
configured.
Since Proxmox VE 6.1 you can apply pending network changes over the web-interface, using the Apply
Configuration button in the Network panel of a node.
To install ifupdown2 ensure you have the latest Proxmox VE updates installed, then
Warning
installing ifupdown2 will remove ifupdown, but as the removal scripts of ifupdown before version
0.8.35+pve1 have a issue where network is fully stopped on removal a you must ensure that you
have a up to date ifupdown package version.
a Introduced with Debian Buster: https://bugs.debian.org/cgi-bin/bugreport.cgi?bug=945877
With that you’re all set. You can also switch back to the ifupdown variant at any time, if you run into issues.
• Ethernet devices: en*, systemd network interface names. This naming scheme is used for new Proxmox
VE installations since version 5.0.
Proxmox VE Administration Guide 30 / 464
• Ethernet devices: eth[N], where 0 ≤ N (eth0, eth1, . . . ) This naming scheme is used for Proxmox VE
hosts which were installed before the 5.0 release. When upgrading to 5.0, the names are kept as-is.
• VLANs: Simply add the VLAN number to the device name, separated by a period (eno1.50, bond1.30)
This makes it easier to debug networks problems, because the device name implies the device type.
Systemd uses the two character prefix en for Ethernet network devices. The next characters depends on the
device driver and the fact which schema matches first.
• enp3s0f1 — is the NIC on pcibus 3 slot 0 and use the NIC function 1.
Depending on your current network organization and your resources you can choose either a bridged, routed,
or masquerading networking setup.
Proxmox VE server in a private LAN, using an external gateway to reach the internet
The Bridged model makes the most sense in this case, and this is also the default mode on new Proxmox
VE installations. Each of your Guest system will have a virtual interface attached to the Proxmox VE bridge.
This is similar in effect to having the Guest network card directly connected to a new switch on your LAN, the
Proxmox VE host playing the role of the switch.
For this setup, you can use either a Bridged or Routed model, depending on what your provider allows.
Proxmox VE Administration Guide 31 / 464
In that case the only way to get outgoing network accesses for your guest systems is to use Masquerading.
For incoming network access to your guests, you will need to configure Port Forwarding.
For further flexibility, you can configure VLANs (IEEE 802.1q) and network bonding, also known as "link
aggregation". That way it is possible to build complex and flexible virtual networks.
eno1 eno1
eno1
Bridges are like physical network switches implemented in software. All virtual guests can share a single
bridge, or you can create multiple bridges to separate network domains. Each host can have up to 4094
bridges.
The installation program creates a single bridge named vmbr0, which is connected to the first Ethernet
card. The corresponding configuration in /etc/network/interfaces might look like this:
auto lo
iface lo inet loopback
auto vmbr0
iface vmbr0 inet static
address 192.168.10.2
netmask 255.255.255.0
gateway 192.168.10.1
Proxmox VE Administration Guide 32 / 464
bridge-ports eno1
bridge-stp off
bridge-fd 0
Virtual machines behave as if they were directly connected to the physical network. The network, in turn,
sees each virtual machine as having its own MAC, even though there is only one network cable connecting
all of these VMs to the network.
Most hosting providers do not support the above setup. For security reasons, they disable networking as
soon as they detect multiple MAC addresses on a single interface.
Tip
Some providers allow you to register additional MACs through their management interface. This avoids
the problem, but can be clumsy to configure because you need to register a MAC for each of your VMs.
You can avoid the problem by “routing” all traffic via a single interface. This makes sure that all network
packets use the same MAC address.
Provider Gateway
198.51.100.1
ip_forward = 1
vmbr0 proxy_arp = 1 eno0
203.0.113.17/28 198.51.100.5/29
tap100i0
Node: proxmox
A common scenario is that you have a public IP (assume 198.51.100.5 for this example), and an addi-
tional IP block for your VMs (203.0.113.16/29). We recommend the following setup for such situations:
auto lo
iface lo inet loopback
Proxmox VE Administration Guide 33 / 464
auto eno1
iface eno1 inet static
address 198.51.100.5
netmask 255.255.255.0
gateway 198.51.100.1
post-up echo 1 > /proc/sys/net/ipv4/ip_forward
post-up echo 1 > /proc/sys/net/ipv4/conf/eno1/proxy_arp
auto vmbr0
iface vmbr0 inet static
address 203.0.113.17
netmask 255.255.255.248
bridge-ports none
bridge-stp off
bridge-fd 0
Masquerading allows guests having only a private IP address to access the network by using the host IP
address for outgoing traffic. Each outgoing packet is rewritten by iptables to appear as originating from
the host, and responses are rewritten accordingly to be routed to the original sender.
auto lo
iface lo inet loopback
auto eno1
#real IP address
iface eno1 inet static
address 198.51.100.5
netmask 255.255.255.0
gateway 198.51.100.1
auto vmbr0
#private sub network
iface vmbr0 inet static
address 10.10.10.1
netmask 255.255.255.0
bridge-ports none
bridge-stp off
bridge-fd 0
Note
In some masquerade setups with firewall enabled, conntrack zones might be needed for outgoing connec-
tions. Otherwise the firewall could block outgoing connections since they will prefer the POSTROUTING
of the VM bridge (and not MASQUERADE).
Bonding (also called NIC teaming or Link Aggregation) is a technique for binding multiple NIC’s to a single
network device. It is possible to achieve different goals, like make the network fault-tolerant, increase the
performance or both together.
High-speed hardware like Fibre Channel and the associated switching hardware can be quite expensive. By
doing link aggregation, two NICs can appear as one logical interface, resulting in double speed. This is a
native Linux kernel feature that is supported by most switches. If your nodes have multiple Ethernet ports,
you can distribute your points of failure by running network cables to different switches and the bonded
connection will failover to one cable or the other in case of network trouble.
Aggregated links can improve live-migration delays and improve the speed of replication of data between
Proxmox VE Cluster nodes.
There are 7 modes for bonding:
• Round-robin (balance-rr): Transmit network packets in sequential order from the first available network
interface (NIC) slave through the last. This mode provides load balancing and fault tolerance.
• Active-backup (active-backup): Only one NIC slave in the bond is active. A different slave becomes
active if, and only if, the active slave fails. The single logical bonded interface’s MAC address is externally
visible on only one NIC (port) to avoid distortion in the network switch. This mode provides fault tolerance.
• XOR (balance-xor): Transmit network packets based on [(source MAC address XOR’d with destination
MAC address) modulo NIC slave count]. This selects the same NIC slave for each destination MAC
address. This mode provides load balancing and fault tolerance.
• Broadcast (broadcast): Transmit network packets on all slave network interfaces. This mode provides
fault tolerance.
• IEEE 802.3ad Dynamic link aggregation (802.3ad)(LACP): Creates aggregation groups that share the
same speed and duplex settings. Utilizes all slave network interfaces in the active aggregator group ac-
cording to the 802.3ad specification.
Proxmox VE Administration Guide 35 / 464
• Adaptive transmit load balancing (balance-tlb): Linux bonding driver mode that does not require any
special network-switch support. The outgoing network packet traffic is distributed according to the current
load (computed relative to the speed) on each network interface slave. Incoming traffic is received by one
currently designated slave network interface. If this receiving slave fails, another slave takes over the MAC
address of the failed receiving slave.
• Adaptive load balancing (balance-alb): Includes balance-tlb plus receive load balancing (rlb) for IPV4
traffic, and does not require any special network switch support. The receive load balancing is achieved by
ARP negotiation. The bonding driver intercepts the ARP Replies sent by the local system on their way out
and overwrites the source hardware address with the unique hardware address of one of the NIC slaves in
the single logical bonded interface such that different network-peers use different MAC addresses for their
network packet traffic.
If your switch support the LACP (IEEE 802.3ad) protocol then we recommend using the corresponding
bonding mode (802.3ad). Otherwise you should generally use the active-backup mode.
If you intend to run your cluster network on the bonding interfaces, then you have to use active-passive mode
on the bonding interfaces, other modes are unsupported.
The following bond configuration can be used as distributed/shared storage network. The benefit would be
that you get more speed and the network will be fault-tolerant.
auto lo
iface lo inet loopback
auto bond0
iface bond0 inet static
bond-slaves eno1 eno2
address 192.168.1.2
netmask 255.255.255.0
bond-miimon 100
bond-mode 802.3ad
bond-xmit-hash-policy layer2+3
auto vmbr0
iface vmbr0 inet static
address 10.10.10.2
netmask 255.255.255.0
gateway 10.10.10.1
bridge-ports eno3
bridge-stp off
bridge-fd 0
Proxmox VE Administration Guide 36 / 464
bond0 bond0
LACP LACP
bond0 bond0
vmbr0 vmbr0
10.10.10.2/24 10.10.10.3/24
tap100i0 tap100i0
ens18 ens18
VM 100 VM 200
10.10.10.100 10.10.10.200
Another possibility it to use the bond directly as bridge port. This can be used to make the guest network
fault-tolerant.
auto lo
iface lo inet loopback
auto bond0
iface bond0 inet manual
bond-slaves eno1 eno2
bond-miimon 100
bond-mode 802.3ad
bond-xmit-hash-policy layer2+3
auto vmbr0
iface vmbr0 inet static
address 10.10.10.2
netmask 255.255.255.0
gateway 10.10.10.1
bridge-ports bond0
bridge-stp off
bridge-fd 0
Proxmox VE Administration Guide 37 / 464
A virtual LAN (VLAN) is a broadcast domain that is partitioned and isolated in the network at layer two. So it
is possible to have multiple networks (4096) in a physical network, each independent of the other ones.
Each VLAN network is identified by a number often called tag. Network packages are then tagged to identify
which virtual network they belong to.
Proxmox VE supports this setup out of the box. You can specify the VLAN tag when you create a VM.
The VLAN tag is part of the guest network configuration. The networking layer supports different modes to
implement VLANs, depending on the bridge configuration:
• VLAN awareness on the Linux bridge: In this case, each guest’s virtual network card is assigned to
a VLAN tag, which is transparently supported by the Linux bridge. Trunk mode is also possible, but that
makes configuration in the guest necessary.
• "traditional" VLAN on the Linux bridge: In contrast to the VLAN awareness method, this method is not
transparent and creates a VLAN device with associated bridge for each VLAN. That is, creating a guest on
VLAN 5 for example, would create two interfaces eno1.5 and vmbr0v5, which would remain until a reboot
occurs.
• Open vSwitch VLAN: This mode uses the OVS VLAN feature.
• Guest configured VLAN: VLANs are assigned inside the guest. In this case, the setup is completely
done inside the guest and can not be influenced from the outside. The benefit is that you can use more
than one VLAN on a single virtual NIC.
To allow host communication with an isolated network. It is possible to apply VLAN tags to any network
device (NIC, Bond, Bridge). In general, you should configure the VLAN on the interface with the least
abstraction layers between itself and the physical NIC.
For example, in a default configuration where you want to place the host management address on a separate
VLAN.
Example: Use VLAN 5 for the Proxmox VE management IP with traditional Linux bridge
auto lo
iface lo inet loopback
auto vmbr0v5
iface vmbr0v5 inet static
address 10.10.10.2
Proxmox VE Administration Guide 38 / 464
netmask 255.255.255.0
gateway 10.10.10.1
bridge-ports eno1.5
bridge-stp off
bridge-fd 0
auto vmbr0
iface vmbr0 inet manual
bridge-ports eno1
bridge-stp off
bridge-fd 0
Example: Use VLAN 5 for the Proxmox VE management IP with VLAN aware Linux bridge
auto lo
iface lo inet loopback
auto vmbr0.5
iface vmbr0.5 inet static
address 10.10.10.2
netmask 255.255.255.0
gateway 10.10.10.1
auto vmbr0
iface vmbr0 inet manual
bridge-ports eno1
bridge-stp off
bridge-fd 0
bridge-vlan-aware yes
The next example is the same setup but a bond is used to make this network fail-safe.
Example: Use VLAN 5 with bond0 for the Proxmox VE management IP with traditional Linux bridge
auto lo
iface lo inet loopback
auto bond0
iface bond0 inet manual
bond-slaves eno1 eno2
bond-miimon 100
bond-mode 802.3ad
bond-xmit-hash-policy layer2+3
Proxmox VE Administration Guide 39 / 464
auto vmbr0v5
iface vmbr0v5 inet static
address 10.10.10.2
netmask 255.255.255.0
gateway 10.10.10.1
bridge-ports bond0.5
bridge-stp off
bridge-fd 0
auto vmbr0
iface vmbr0 inet manual
bridge-ports bond0
bridge-stp off
bridge-fd 0
The Proxmox VE cluster stack itself relies heavily on the fact that all the nodes have precisely synchronized
time. Some other components, like Ceph, also refuse to work properly if the local time on nodes is not in
sync.
Time synchronization between nodes can be achieved with the “Network Time Protocol” (NTP). Proxmox VE
uses systemd-timesyncd as NTP client by default, preconfigured to use a set of public servers. This
setup works out of the box in most cases.
In some cases, it might be desired to not use the default NTP servers. For example, if your Proxmox VE
nodes do not have access to the public internet (e.g., because of restrictive firewall rules), you need to setup
local NTP servers and tell systemd-timesyncd to use them:
File /etc/systemd/timesyncd.conf
[Time]
NTP=ntp1.example.com ntp2.example.com ntp3.example.com ntp4.example.com
After restarting the synchronization service (systemctl restart systemd-timesyncd) you should
verify that your newly configured NTP servers are used by checking the journal (journalctl --since
-1h -u systemd-timesyncd):
...
Oct 07 14:58:36 node1 systemd[1]: Stopping Network Time Synchronization...
Oct 07 14:58:36 node1 systemd[1]: Starting Network Time Synchronization...
Oct 07 14:58:36 node1 systemd[1]: Started Network Time Synchronization.
Proxmox VE Administration Guide 40 / 464
In Proxmox VE, you can define external metric servers, which will periodically receive various stats about
your hosts, virtual guests and storages.
Currently supported are:
The external metric server definitions are saved in /etc/pve/status.cfg, and can be edited through the web
interface.
Proxmox VE Administration Guide 41 / 464
The default port is set to 2003 and the default graphite path is proxmox.
By default, Proxmox VE sends the data over UDP, so the graphite server has to be configured to accept this.
Here the maximum transmission unit (MTU) can be configured for environments not using the standard 1500
MTU.
You can also configure the plugin to use TCP. In order not to block the important pvestatd statistic collec-
tion daemon, a timeout is required to cope with network problems.
Proxmox VE sends the data over UDP, so the influxdb server has to be configured for this. The MTU can
also be configured here, if necessary.
Here is an example configuration for influxdb (on your influxdb server):
[[udp]]
enabled = true
bind-address = "0.0.0.0:8089"
database = "proxmox"
batch-size = 1000
batch-timeout = "1s"
With this configuration, your server listens on all IP addresses on port 8089, and writes the data in the
proxmox database
Although a robust and redundant storage is recommended, it can be very helpful to monitor the health of
your local disks.
Starting with Proxmox VE 4.3, the package smartmontools 1 is installed and required. This is a set of tools
to monitor and control the S.M.A.R.T. system for local hard disks.
1 smartmontools homepage https://www.smartmontools.org
Proxmox VE Administration Guide 42 / 464
You can get the status of a disk by issuing the following command:
# smartctl -a /dev/sdX
# smartctl -s on /dev/sdX
For more information on how to use smartctl, please see man smartctl.
By default, smartmontools daemon smartd is active and enabled, and scans the disks under /dev/sdX and
/dev/hdX every 30 minutes for errors and warnings, and sends an e-mail to root if it detects a problem.
For more information about how to configure smartd, please see man smartd and man smartd.conf.
If you use your hard disks with a hardware raid controller, there are most likely tools to monitor the disks in
the raid array and the array itself. For more information about this, please refer to the vendor of your raid
controller.
Most people install Proxmox VE directly on a local disk. The Proxmox VE installation CD offers several
options for local disk management, and the current default setup uses LVM. The installer let you select a
single disk for such setup, and uses that disk as physical volume for the Volume Group (VG) pve. The
following output is from a test installation using a small 8GB disk:
# pvs
PV VG Fmt Attr PSize PFree
/dev/sda3 pve lvm2 a-- 7.87g 876.00m
# vgs
VG #PV #LV #SN Attr VSize VFree
pve 1 3 0 wz--n- 7.87g 876.00m
The installer allocates three Logical Volumes (LV) inside this VG:
# lvs
LV VG Attr LSize Pool Origin Data% Meta%
data pve twi-a-tz-- 4.38g 0.00 0.63
root pve -wi-ao---- 1.75g
swap pve -wi-ao---- 896.00m
Proxmox VE Administration Guide 43 / 464
root
Formatted as ext4, and contains the operation system.
swap
Swap partition
data
This volume uses LVM-thin, and is used to store VM images. LVM-thin is preferable for this task,
because it offers efficient support for snapshots and clones.
For Proxmox VE versions up to 4.1, the installer creates a standard logical volume called “data”, which is
mounted at /var/lib/vz.
Starting from version 4.2, the logical volume “data” is a LVM-thin pool, used to store block based guest
images, and /var/lib/vz is simply a directory on the root file system.
3.7.1 Hardware
We highly recommend to use a hardware RAID controller (with BBU) for such setups. This increases perfor-
mance, provides redundancy, and make disk replacements easier (hot-pluggable).
LVM itself does not need any special hardware, and memory requirements are very low.
3.7.2 Bootloader
We install two boot loaders by default. The first partition contains the standard GRUB boot loader. The
second partition is an EFI System Partition (ESP), which makes it possible to boot on EFI systems.
Let’s assume we have an empty disk /dev/sdb, onto which we want to create a volume group named
“vmdata”.
Caution
Please note that the following commands will destroy all existing data on /dev/sdb.
# sgdisk -N 1 /dev/sdb
# mkfs.ext4 /dev/pve/vz
Warning
be sure that /var/lib/vz is empty. On a default installation it’s not.
Resize the LV and the metadata pool can be achieved with the following command.
Note
When extending the data pool, the metadata pool must also be extended.
Proxmox VE Administration Guide 45 / 464
A thin pool has to be created on top of a volume group. How to create a volume group see Section LVM.
ZFS is a combined file system and logical volume manager designed by Sun Microsystems. Starting with
Proxmox VE 3.4, the native Linux kernel port of the ZFS file system is introduced as optional file system and
also as an additional selection for the root file system. There is no need for manually compile ZFS modules
- all packages are included.
By using ZFS, its possible to achieve maximum enterprise features with low budget hardware, but also high
performance systems by leveraging SSD caching or even SSD only setups. ZFS can replace cost intense
hardware raid cards by moderate CPU and memory load combined with easy management.
G ENERAL ZFS ADVANTAGES
• Reliable
• Snapshots
• Copy-on-write clone
• Various raid levels: RAID0, RAID1, RAID10, RAIDZ-1, RAIDZ-2 and RAIDZ-3
• Self healing
• Open Source
• Encryption
• ...
Proxmox VE Administration Guide 46 / 464
3.8.1 Hardware
ZFS depends heavily on memory, so you need at least 8GB to start. In practice, use as much you can get
for your hardware/budget. To prevent data corruption, we recommend the use of high quality ECC RAM.
If you use a dedicated cache and/or log disk, you should use an enterprise class SSD (e.g. Intel SSD DC
S3700 Series). This can increase the overall performance significantly.
Important
Do not use ZFS on top of hardware controller which has its own cache management. ZFS needs to
directly communicate with disks. An HBA adapter is the way to go, or something like LSI controller
flashed in “IT” mode.
If you are experimenting with an installation of Proxmox VE inside a VM (Nested Virtualization), don’t use
virtio for disks of that VM, since they are not supported by ZFS. Use IDE or SCSI instead (works also
with virtio SCSI controller type).
When you install using the Proxmox VE installer, you can choose ZFS for the root file system. You need to
select the RAID type at installation time:
RAID0 Also called “striping”. The capacity of such volume is the sum of the capacities of all
disks. But RAID0 does not add any redundancy, so the failure of a single drive
makes the volume unusable.
RAID1 Also called “mirroring”. Data is written identically to all disks. This mode requires at
least 2 disks with the same size. The resulting capacity is that of a single disk.
The installer automatically partitions the disks, creates a ZFS pool called rpool, and installs the root file
system on the ZFS subvolume rpool/ROOT/pve-1.
Another subvolume called rpool/data is created to store VM images. In order to use that with the
Proxmox VE tools, the installer creates the following configuration entry in /etc/pve/storage.cfg:
zfspool: local-zfs
pool rpool/data
sparse
content images,rootdir
Proxmox VE Administration Guide 47 / 464
After installation, you can view your ZFS pool status using the zpool command:
# zpool status
pool: rpool
state: ONLINE
scan: none requested
config:
The zfs command is used configure and manage your ZFS file systems. The following command lists all
file systems after installation:
# zfs list
NAME USED AVAIL REFER MOUNTPOINT
rpool 4.94G 7.68T 96K /rpool
rpool/ROOT 702M 7.68T 96K /rpool/ROOT
rpool/ROOT/pve-1 702M 7.68T 702M /
rpool/data 96K 7.68T 96K /rpool/data
rpool/swap 4.25G 7.69T 64K -
There are a few factors to take into consideration when choosing the layout of a ZFS pool. The basic building
block of a ZFS pool is the virtual device, or vdev. All vdevs in a pool are used equally and the data is striped
among them (RAID0). Check the zpool(8) manpage for more details on vdevs.
Performance
Each vdev type has different performance behaviors. The two parameters of interest are the IOPS (In-
put/Output Operations per Second) and the bandwidth with which data can be written or read.
A mirror vdev (RAID1) will approximately behave like a single disk in regards to both parameters when writing
data. When reading data if will behave like the number of disks in the mirror.
A common situation is to have 4 disks. When setting it up as 2 mirror vdevs (RAID10) the pool will have
the write characteristics as two single disks in regard of IOPS and bandwidth. For read operations it will
resemble 4 single disks.
A RAIDZ of any redundancy level will approximately behave like a single disk in regard of IOPS with a lot of
bandwidth. How much bandwidth depends on the size of the RAIDZ vdev and the redundancy level.
Proxmox VE Administration Guide 48 / 464
For running VMs, IOPS is the more important metric in most situations.
While a pool made of mirror vdevs will have the best performance characteristics, the usable space will be
50% of the disks available. Less if a mirror vdev consists of more than 2 disks, for example in a 3-way mirror.
At least one healthy disk per mirror is needed for the pool to stay functional.
The usable space of a RAIDZ type vdev of N disks is roughly N-P, with P being the RAIDZ-level. The RAIDZ-
level indicates how many arbitrary disks can fail without losing data. A special case is a 4 disk pool with
RAIDZ2. In this situation it is usually better to use 2 mirror vdevs for the better performance as the usable
space will be the same.
Another important factor when using any RAIDZ level is how ZVOL datasets, which are used for VM disks,
behave. For each data block the pool needs parity data which is at least the size of the minimum block size
defined by the ashift value of the pool. With an ashift of 12 the block size of the pool is 4k. The default
block size for a ZVOL is 8k. Therefore, in a RAIDZ2 each 8k block written will cause two additional 4k parity
blocks to be written, 8k + 4k + 4k = 16k. This is of course a simplified approach and the real situation will be
slightly different with metadata, compression and such not being accounted for in this example.
This behavior can be observed when checking the following properties of the ZVOL:
• volsize
• used (if the pool is thin provisioned and without snapshots present)
volsize is the size of the disk as it is presented to the VM, while refreservation shows the reserved
space on the pool which includes the expected space needed for the parity data. If the pool is thin provi-
sioned, the refreservation will be set to 0. Another way to observe the behavior is to compare the
used disk space within the VM and the used property. Be aware that snapshots will skew the value.
There are a few options to counter the increased use of space:
The volblocksize property can only be set when creating a ZVOL. The default value can be changed in
the storage configuration. When doing this, the guest needs to be tuned accordingly and depending on the
use case, the problem of write amplification if just moved from the ZFS layer up to the guest.
Using ashift=9 when creating the pool can lead to bad performance, depending on the disks underneath,
and cannot be changed later on.
Mirror vdevs (RAID1, RAID10) have favorable behavior for VM workloads. Use them, unless your environ-
ment has specific needs and characteristics where RAIDZ performance characteristics are acceptable.
Proxmox VE Administration Guide 49 / 464
3.8.4 Bootloader
Depending on whether the system is booted in EFI or legacy BIOS mode the Proxmox VE installer sets
up either grub or systemd-boot as main bootloader. See the chapter on Proxmox VE host bootladers
Section 3.11 for details.
This section gives you some usage examples for common tasks. ZFS itself is really powerful and provides
many options. The main commands to manage ZFS are zfs and zpool. Both commands come with great
manual pages, which can be read with:
# man zpool
# man zfs
To create a new pool, at least one disk is needed. The ashift should have the same sector-size (2 power
of ashift) or larger as the underlying disk.
Minimum 1 disk
Minimum 2 disks
Minimum 4 disks
Minimum 3 disks
Minimum 4 disks
It is possible to use a dedicated cache drive partition to increase the performance (use SSD).
As <device> it is possible to use more devices, like it’s shown in "Create a new pool with RAID*".
If you have a pool without cache and log. First partition the SSD in 2 partition with parted or gdisk
Important
Always use GPT partition tables.
The maximum size of a log device should be about half the size of physical memory, so this is usually quite
small. The rest of the SSD can be used as cache.
Depending on how Proxmox VE was installed it is either using grub or systemd-boot as bootloader
(see Host Bootloader Section 3.11).
The first steps of copying the partition table, reissuing GUIDs and replacing the ZFS partition are the same.
To make the system bootable from the new disk, different steps are needed which depend on the bootloader
in use.
Note
Use the zpool status -v command to monitor how far the resilvering process of the new disk has
progressed.
With systemd-boot:
Note
ESP stands for EFI System Partition, which is setup as partition #2 on bootable disks setup by the Proxmox
VE installer since version 5.4. For details, see Setting up a new partition for use as synced ESP Setting
up a new partition for use as synced ESP.
Proxmox VE Administration Guide 52 / 464
With grub:
ZFS comes with an event daemon, which monitors events generated by the ZFS kernel module. The daemon
can also send emails on ZFS events like pool errors. Newer ZFS packages ship the daemon in a separate
package, and you can install it using apt-get:
To activate the daemon it is necessary to edit /etc/zfs/zed.d/zed.rc with your favourite editor, and
uncomment the ZED_EMAIL_ADDR setting:
ZED_EMAIL_ADDR="root"
Please note Proxmox VE forwards mails to root to the email address configured for the root user.
Important
The only setting that is required is ZED_EMAIL_ADDR. All other settings are optional.
It is good to use at most 50 percent (which is the default) of the system memory for ZFS ARC to prevent per-
formance shortage of the host. Use your preferred editor to change the configuration in /etc/modprobe.d/zfs
and insert:
Important
If your root file system is ZFS you must update your initramfs every time this value changes:
# update-initramfs -u
Proxmox VE Administration Guide 53 / 464
Swap-space created on a zvol may generate some troubles, like blocking the server or generating a high IO
load, often seen when starting a Backup to an external Storage.
We strongly recommend to use enough memory, so that you normally do not run into low memory situations.
Should you need or want to add swap, it is preferred to create a partition on a physical disk and use it
as swapdevice. You can leave some space free for this purpose in the advanced options of the installer.
Additionally, you can lower the “swappiness” value. A good value for servers is 10:
# sysctl -w vm.swappiness=10
To make the swappiness persistent, open /etc/sysctl.conf with an editor of your choice and add the
following line:
vm.swappiness = 10
Value Strategy
vm.swappiness = 0 The kernel will swap only to avoid an out of memory condition
vm.swappiness = 1 Minimum amount of swapping without disabling it entirely.
vm.swappiness = 10 This value is sometimes recommended to improve performance
when sufficient memory exists in a system.
vm.swappiness = 60 The default value.
vm.swappiness = 100 The kernel will swap aggressively.
ZFS on Linux version 0.8.0 introduced support for native encryption of datasets. After an upgrade from
previous ZFS on Linux versions, the encryption feature can be enabled per pool:
Warning
There is currently no support for booting from pools with encrypted datasets using Grub, and only
limited support for automatically unlocking encrypted datasets on boot. Older versions of ZFS
without encryption support will not be able to decrypt stored data.
Note
It is recommended to either unlock storage datasets manually after booting, or to write a custom unit to
pass the key material needed for unlocking on boot to zfs load-key.
Warning
Establish and test a backup procedure before enabling encryption of production data. If the as-
sociated key material/passphrase/keyfile has been lost, accessing the encrypted data is no longer
possible.
Encryption needs to be setup when creating datasets/zvols, and is inherited by default to child datasets. For
example, to create an encrypted dataset tank/encrypted_data and configure it as storage in Proxmox
VE, run the following commands:
All guest volumes/disks create on this storage will be encrypted with the shared key material of the parent
dataset.
To actually use the storage, the associated key material needs to be loaded and the dataset needs to be
mounted. This can be done in one step with:
It is also possible to use a (random) keyfile instead of prompting for a passphrase by setting the keylocation
and keyformat properties, either at creation time or with zfs change-key on existing datasets:
Warning
When using a keyfile, special care needs to be taken to secure the keyfile against unauthorized
access or accidental loss. Without the keyfile, it is not possible to access the plaintext data!
Proxmox VE Administration Guide 55 / 464
A guest volume created underneath an encrypted dataset will have its encryptionroot property set
accordingly. The key material only needs to be loaded once per encryptionroot to be available to all encrypted
datasets underneath it.
See the encryptionroot, encryption, keylocation, keyformat and keystatus proper-
ties, the zfs load-key, zfs unload-key and zfs change-key commands and the Encryption
section from man zfs for more details and advanced usage.
When compression is enabled on a dataset, ZFS tries to compress all new blocks before writing them and
decompresses them on reading. Already existing data will not be compressed retroactively.
You can enable compression with:
We recommend using the lz4 algorithm, because it adds very little CPU overhead. Other algorithms like
lzjb and gzip-N, where N is an integer from 1 (fastest) to 9 (best compression ratio), are also avail-
able. Depending on the algorithm and how compressible the data is, having compression enabled can even
increase I/O performance.
You can disable compression at any time with:
Since version 0.8.0 ZFS supports special devices. A special device in a pool is used to store meta-
data, deduplication tables, and optionally small file blocks.
A special device can improve the speed of a pool consisting of slow spinning hard disks with a lot of
metadata changes. For example workloads that involve creating, updating or deleting a large number of
files will benefit from the presence of a special device. ZFS datasets can also be configured to store
whole small files on the special device which can further improve the performance. Use fast SSDs for the
special device.
Important
The redundancy of the special device should match the one of the pool, since the special
device is a point of failure for the whole pool.
Warning
Adding a special device to a pool cannot be undone!
Proxmox VE Administration Guide 56 / 464
Important
If the value for special_small_blocks is greater than or equal to the recordsize (default
128K) of the dataset, all data will be written to the special device, so be careful!
Setting the special_small_blocks property on a pool will change the default value of that property
for all child ZFS datasets (for example all containers in the pool will opt in for small file blocks).
The Proxmox VE node management tool (pvenode) allows to control node specific settings and resources.
Currently pvenode allows to set a node’s description and to manage the node’s SSL certificates used for
the API and the web GUI through pveproxy.
Proxmox VE Administration Guide 57 / 464
3.9.1 Wake-on-LAN
Wake-on-LAN (WoL) allows to switch on a sleeping computer in the network by sending a magic packet.
At least one NIC must support this feature and the respective option needs to be enabled in the computers
firmware (BIOS/UEFI) configuration. The option name can vary from Enable Wake-on-Lan to Power On By
PCIE Device, check your motherboards vendor manual, if unsure. ethtool can be used to check the WoL
configuration of <interface> by running:
pvenode allows to wake sleeping members of a cluster via WoL using the command:
This broadcasts the WoL magic packet on UDP port 9, containing the MAC address of <node> obtained
from the wakeonlan property. The node specific wakeonlan property can be set by the following com-
mand:
Each Proxmox VE cluster creates by default its own (self-signed) Certificate Authority (CA) and generates
a certificate for each node which gets signed by the aforementioned CA. These certificates are used for
encrypted communication with the cluster’s pveproxy service and the Shell/Console feature if SPICE is
used.
The CA certificate and key are stored in the Proxmox Cluster File System (pmxcfs) Chapter 6.
The REST API and web GUI are provided by the pveproxy service, which runs on each node.
You have the following options for the certificate used by pveproxy:
3. use ACME (Let’s Encrypt) to get a trusted certificate with automatic renewal, this is also integrated in
the Proxmox VE API and Webinterface.
Proxmox VE Administration Guide 58 / 464
Note
Keep in mind that /etc/pve/local is a node specific symlink to /etc/pve/nodes/NODENAME.
Certificates are managed with the Proxmox VE Node management command (see the pvenode(1) man-
page).
Warning
Do not replace or manually modify the automatically generated node certificate files in
/etc/pve/local/pve-ssl.pem and /etc/pve/local/pve-ssl.key or the cluster
CA files in /etc/pve/pve-root-ca.pem and /etc/pve/priv/pve-root-ca.key.
If you already have a certificate which you want to use for a Proxmox VE node you can upload that certificate
simply over the web interface.
Note that the certificates key file, if provided, mustn’t be password protected.
Proxmox VE includes an implementation of the Automatic Certificate Management Environment ACME pro-
tocol, allowing Proxmox VE admins to interface with Let’s Encrypt for easy setup of trusted TLS certificates
which are accepted out of the box on most modern operating systems and browsers.
Currently the two ACME endpoints implemented are the Let’s Encrypt (LE) production and its staging en-
vironment. Our ACME client supports validation of http-01 challenges using a built-in webserver and
validation of dns-01 challenges using a DNS plugin supporting all the DNS API endpoints acme.sh does.
Proxmox VE Administration Guide 59 / 464
ACME Account
You need to register an ACME account per cluster with the endpoint you want to use. The email address
used for that account will server as contact point for renewal-due or similar notifications from the ACME
endpoint.
You can register and deactivate ACME accounts over the web interface Datacenter -> ACME or using
the pvenode command line tool.
Tip
Because of rate-limits you should use LE staging for experiments or if you use ACME for the first time.
ACME Plugins
The ACME plugins task is to provide automatic verification that you, and thus the Proxmox VE cluster under
your operation, are the real owner of a domain. This is the basis building block for automatic certificate
management.
The ACME protocol specifies different types of challenges, for example the http-01 where a webserver
provides a file with a certain value to prove that it controls a domain. Sometimes this isn’t possible, either
because of technical limitations or if the address a domain points to is not reachable from the public internet.
For such cases, one could use the dns-01 challenge. This challenge also provides a certain value, but
through a DNS record on the authority name server of the domain, rather than over a text file.
Proxmox VE Administration Guide 60 / 464
Proxmox VE supports both of those challenge types out of the box, you can configure plugins either over the
web interface under Datacenter -> ACME, or using the pvenode acme plugin add command.
ACME Plugin configurations are stored in /etc/pve/priv/acme/plugins.cfg. A plugin is available
for all nodes in the cluster.
Node Domains
Each domain is node specific. You can add new or manage existing domain entries under Node ->
Certificates, or using the pvenode config command.
After configuring the desired domain(s) for a node and ensuring that the desired ACME account is selected,
you can order your new certificate over the web-interface. On success the interface will reload after 10
seconds.
Renewal will happen automatically Section 3.10.7.
There is always an implicitly configured standalone plugin for validating http-01 challenges via the
built-in webserver spawned on port 80.
Proxmox VE Administration Guide 61 / 464
Note
The name standalone means that it can provide the validation on it’s own, without any third party
service. So, this plugin works also for cluster nodes.
There are a few prerequisites to use it for certificate management with Let’s Encrypts ACME.
On systems where external access for validation via the http-01 method is not possible or desired, it is
possible to use the dns-01 validation method. This validation method requires a DNS server that allows
provisioning of TXT records via an API.
Proxmox VE re-uses the DNS plugins developed for the acme.sh 2 project, please refer to its documenta-
tion for details on configuration of specific APIs.
The easiest way to configure a new plugin with the DNS API is using the web interface (Datacenter ->
ACME).
Choose DNS as challenge type. Then you can select your API provider, enter the credential data to access
your account over their API.
Tip
See the acme.sh How to use DNS API wiki for more detailed information about getting API credentials for
your provider.
As there are so many API endpoints Proxmox VE autogenerates the form for the credentials, but not all
providers are annotated yet. For those you will see a bigger text area, simply copy all the credentials
KEY=VALUE pairs in there.
2 acme.sh https://github.com/acmesh-official/acme.sh
Proxmox VE Administration Guide 62 / 464
A special alias mode can be used to handle the validation on a different domain/DNS server, in case your
primary/real DNS does not support provisioning via an API. Manually set up a permanent CNAME record for
_acme-challenge.domain1.example pointing to _acme-challenge.domain2.example and
set the alias property in the Proxmox VE node configuration file to domain2.example to allow the DNS
server of domain2.example to validate all challenges for domain1.example.
Combination of Plugins
Combining http-01 and dns-01 validation is possible in case your node is reachable via multiple do-
mains with different requirements / DNS provisioning capabilities. Mixing DNS APIs from multiple providers
or instances is also possible by specifying different plugin instances per domain.
Tip
Accessing the same service over multiple domains increases complexity and should be avoided if possible.
If a node has been successfully configured with an ACME-provided certificate (either via pvenode or via
the GUI), the certificate will be automatically renewed by the pve-daily-update.service. Currently,
renewal will be attempted if the certificate has expired already, or will expire in the next 30 days.
Note
the account registration steps are the same no matter which plugins are used, and are not repeated here.
Note
OVH_AK and OVH_AS need to be obtained from OVH according to the OVH API documentation
First you need to get all information so you and Proxmox VE can access the API.
(open validation URL and follow instructions to link Application Key with ←-
account/Consumer Key)
root@proxmox:~# pvenode acme plugin add dns example_plugin --api ovh --data ←-
/path/to/api_token
root@proxmox:~# pvenode acme plugin config example_plugin
┌────────┬&#x
│ key │ value │
╞════════╪&#x
│ api │ ovh │
├────────┼&#x
│ data │ OVH_AK=XXXXXXXXXXXXXXXX │
│ │ OVH_AS=YYYYYYYYYYYYYYYYYYYYYYYYYYYYYYYY │
│ │ OVH_CK=ZZZZZZZZZZZZZZZZZZZZZZZZZZZZZZZZ │
├────────┼&#x
│ digest │ 867fcf556363ca1bea866863093fcab83edf47a1 │
├────────┼&#x
│ plugin │ example_plugin │
├────────┼&#x
│ type │ dns │
└────────┴&#x
At last you can configure the domain you want to get certitficates for and place the certificate order for it:
Creating CSR
Proxmox VE Administration Guide 65 / 464
Downloading certificate
Setting pveproxy certificate and key
Restarting pveproxy
Task OK
Changing the ACME directory for an account is unsupported, but as Proxmox VE supports more than one
account you can just create a new one with the production (trusted) ACME directory as endpoint. You can
also deactivate the staging account and recreate it.
Example: Changing the default ACME account from staging to directory using pvenode
Proxmox VE currently uses one of two bootloaders depending on the disk setup selected in the installer.
For EFI Systems installed with ZFS as the root filesystem systemd-boot is used. All other deployments
use the standard grub bootloader (this usually also applies to systems which are installed on top of Debian).
The Proxmox VE installer creates 3 partitions on the bootable disks selected for installation. The bootable
disks are:
Proxmox VE Administration Guide 66 / 464
• a third partition spanning the set hdsize parameter or the remaining space used for the chosen storage
type
grub in BIOS mode (--target i386-pc) is installed onto the BIOS Boot Partition of all bootable disks
for supporting older systems.
The simplest and most reliable way to determine which bootloader is used, is to watch the boot process of
the Proxmox VE node.
You will either see the blue box of grub or the simple black on white systemd-boot.
Determining the bootloader from a running system might not be 100% accurate. The safest way is to run the
following command:
# efibootmgr -v
If it returns a message that EFI variables are not supported, grub is used in BIOS/Legacy mode.
If the output contains a line that looks similar to the following, grub is used in UEFI mode.
3.11.3 Grub
grub has been the de-facto standard for booting Linux systems for many years and is quite well documented
3.
The kernel and initrd images are taken from /boot and its configuration file /boot/grub/grub.cfg
gets updated by the kernel installation process.
Configuration
Changes to the grub configuration are done via the defaults file /etc/default/grub or config snip-
pets in /etc/default/grub.d. To regenerate the /boot/grub/grub.cfg after a change to the
configuration run:
‘update-grub‘.
3.11.4 Systemd-boot
systemd-boot is a lightweight EFI bootloader. It reads the kernel and initrd images directly from the EFI
Service Partition (ESP) where it is installed. The main advantage of directly loading the kernel from the ESP
is that it does not need to reimplement the drivers for accessing the storage. In the context of ZFS as root
filesystem this means that you can use all optional features on your root pool instead of the subset which is
also present in the ZFS implementation in grub or having to create a separate small boot-pool 4 .
In setups with redundancy (RAID1, RAID10, RAIDZ*) all bootable disks (those being part of the first vdev)
are partitioned with an ESP. This ensures the system boots even if the first boot device fails. The ESPs are
kept in sync by a kernel postinstall hook script /etc/kernel/postinst.d/zz-pve-efiboot. The
script copies certain kernel versions and the initrd images to EFI/proxmox/ on the root of each ESP and
creates the appropriate config files in loader/entries/proxmox-*.conf. The pve-efiboot-tool
script assists in managing both the synced ESPs themselves and their contents.
The following kernel versions are configured by default:
• the latest version of the second-to-last kernel series (e.g. 4.15, 5.0), if applicable
The ESPs are not kept mounted during regular operation, in contrast to grub, which keeps an ESP mounted
on /boot/efi. This helps to prevent filesystem corruption to the vfat formatted ESPs in case of a
system crash, and removes the need to manually adapt /etc/fstab in case the primary boot device fails.
3 Grub Manual https://www.gnu.org/software/grub/manual/grub/grub.html
4 Booting ZFS on root with grub https://github.com/zfsonlinux/zfs/wiki/Debian-Stretch-Root-on-ZFS
Proxmox VE Administration Guide 69 / 464
Configuration
systemd-boot is configured via the file loader/loader.conf in the root directory of an EFI System
Partition (ESP). See the loader.conf(5) manpage for details.
Each bootloader entry is placed in a file of its own in the directory loader/entries/
An example entry.conf looks like this (/ refers to the root of the ESP):
title Proxmox
version 5.0.15-1-pve
options root=ZFS=rpool/ROOT/pve-1 boot=zfs
linux /EFI/proxmox/5.0.15-1-pve/vmlinuz-5.0.15-1-pve
initrd /EFI/proxmox/5.0.15-1-pve/initrd.img-5.0.15-1-pve
Should you wish to add a certain kernel and initrd image to the list of bootable kernels use pve-efiboot-tool
kernel add.
For example run the following to add the kernel with ABI version 5.0.15-1-pve to the list of kernels to
keep installed and synced to all ESPs:
pve-efiboot-tool kernel list will list all kernel versions currently selected for booting:
Run pve-efiboot-tool remove to remove a kernel from the list of manually selected kernels, for
example:
Note
It’s required to run pve-efiboot-tool refresh to update all EFI System Partitions (ESPs) after a
manual kernel addition or removal from above.
Proxmox VE Administration Guide 70 / 464
To format and initialize a partition as synced ESP, e.g., after replacing a failed vdev in an rpool, or when con-
verting an existing system that pre-dates the sync mechanism, pve-efiboot-tool from pve-kernel-help
can be used.
Warning
the format command will format the <partition>, make sure to pass in the right device/par-
tition!
For example, to format an empty partition /dev/sda2 as ESP, run the following:
To setup an existing, unmounted ESP located on /dev/sda2 for inclusion in Proxmox VE’s kernel update
synchronization mechanism, use the following:
Afterwards /etc/kernel/pve-efiboot-uuids should contain a new line with the UUID of the newly
added partition. The init command will also automatically trigger a refresh of all configured ESPs.
To copy and configure all bootable kernels and keep all ESPs listed in /etc/kernel/pve-efiboot-uuids
in sync you just need to run:
pve-efiboot-tool refresh
Note
Both update-initramfs and apt (when necessary) will automatically trigger a refresh.
You can modify the kernel commandline in the following places, depending on the bootloader used:
Grub
The kernel commandline needs to be placed in the variable GRUB_CMDLINE_LINUX_DEFAULT in the file
/etc/default/grub. Running update-grub appends its content to all linux entries in /boot/grub/g
Proxmox VE Administration Guide 71 / 464
Systemd-boot
The kernel commandline needs to be placed as one line in /etc/kernel/cmdline. To apply your
changes, run pve-efiboot-tool refresh, which sets it as the option line for all config files in
loader/entries/proxmox-*.conf.
Proxmox VE Administration Guide 72 / 464
Chapter 4
Proxmox VE is simple. There is no need to install a separate management tool, and everything can be done
through your web browser (Latest Firefox or Google Chrome is preferred). A built-in HTML5 console is used
to access the guest console. As an alternative, SPICE can be used.
Because we use the Proxmox cluster file system (pmxcfs), you can connect to any node to manage the
entire cluster. Each node can manage the entire cluster. There is no need for a dedicated manager node.
You can use the web-based administration interface with any modern browser. When Proxmox VE detects
that you are connecting from a mobile device, you are redirected to a simpler, touch-based user interface.
The web interface can be reached via https://youripaddress:8006 (default login is: root, and the password is
specified during the installation process).
4.1 Features
• Secure access to all Virtual Machines and Containers via SSL encryption (https)
• Fast search-driven interface, capable of handling hundreds and probably thousands of VMs
• Role based permission management for all objects (VMs, storages, nodes, etc.)
4.2 Login
When you connect to the server, you will first see the login window. Proxmox VE supports various authen-
tication backends (Realm), and you can select the language here. The GUI is translated to more than 20
languages.
Note
You can save the user name on the client side by selecting the checkbox at the bottom. This saves some
typing when you login next time.
Header On top. Shows status information and contains buttons for most important actions.
Resource Tree At the left side. A navigation tree where you can select specific objects.
Content Panel Center region. Selected objects display configuration options and status here.
Log Panel At the bottom. Displays log entries for recent tasks. You can double-click on those
log entries to get more details, or to abort a running task.
Note
You can shrink and expand the size of the resource tree and log panel, or completely hide the log panel.
This can be helpful when you work on small displays and want more space to view other content.
4.3.1 Header
On the top left side, the first thing you see is the Proxmox logo. Next to it is the current running version of
Proxmox VE. In the search bar nearside you can search for specific objects (VMs, containers, nodes, . . . ).
This is sometimes faster than selecting an object in the resource tree.
To the right of the search bar we see the identity (login name). The gear symbol is a button opening the
My Settings dialog. There you can customize some client side user interface setting (reset the saved login
name, reset saved layout).
The rightmost part of the header contains four buttons:
4.3.2 My Settings
The My Settings window allows you to set locally stored settings. These include the Dashboard Storages
which allow you to enable or disable specific storages to be counted towards the total amount visible in the
datacenter summary. If no storage is checked the total is the sum of all storages, same as enabling every
single one.
Below the dashboard settings you find the stored user name and a button to clear it as well as a button to
reset every layout in the GUI to its default.
On the right side there are xterm.js Settings. These contain the following options:
This is the main navigation tree. On top of the tree you can select some predefined views, which changes
the structure of the tree below. The default view is Server View, and it shows the following object types:
Node Represents the hosts inside a cluster, where the guests runs.
The main purpose of the log panel is to show you what is currently going on in your cluster. Actions like
creating an new VM are executed in background, and we call such background job a task.
Any output from such task is saved into a separate log file. You can view that log by simply double-click a
task log entry. It is also possible to abort a running task there.
Please note that we display most recent tasks from all cluster nodes here. So you can see when somebody
else is working on another cluster node in real-time.
Note
We remove older and finished task from the log panel to keep that list short. But you can still find those
tasks in the Task History within the node panel.
Some short running actions simply sends logs to all cluster members. You can see those messages in the
Cluster log panel.
When you select something in the resource tree, the corresponding object displays configuration and status
information in the content panel. The following sections give a brief overview of the functionality. Please refer
to the individual chapters inside the reference documentation to get more detailed information.
Proxmox VE Administration Guide 77 / 464
4.4.1 Datacenter
On the datacenter level you can access cluster wide settings and information.
• Search: it is possible to search anything in cluster ,this can be a node, VM, Container, Storage or a pool.
• Options: can show and set defaults, which apply cluster wide.
• Backup: has the capability to schedule Backups. This is cluster wide, so you do not care about where the
VM/Container are on your cluster at schedule time.
• Permissions: will manage user and group permission, LDAP, MS-AD and Two-Factor authentication can
be setup here.
• Firewall: on this level the Proxmox Firewall works cluster wide and makes templates which are cluster
wide available.
• Support: here you get all information about your support subscription.
If you like to have more information about this see the corresponding chapter.
4.4.2 Nodes
• Search: it is possible to search anything on the node, this can be a VM, Container, Storage or a pool.
• System: is for configuring the network, DNS and time, and also shows your syslog.
• Updates: will upgrade the system and inform you about new packages.
• Disks: gives you a brief overview about you physical hard drives and how they are used.
• Ceph: is only used if you have installed a Ceph server on your host. Then you can manage your Ceph
cluster and see the status of it here.
• Subscription: here you can upload you subscription key and get a system overview in case of a support
case.
4.4.3 Guests
There are two different kinds of guests and both can be converted to a template. One of them is a Kernel-
based Virtual Machine (KVM) and the other one a Linux Container (LXC). Generally the navigation is the
same, only some options are different.
In the main management center the VM navigation begins if a VM is selected in the left tree.
The top header contains important VM operation commands like Start, Shutdown, Reset, Remove, Migrate,
Console and Help. Some of them have hidden buttons like Shutdown has Stop and Console contains the
different console types SPICE, noVNC and xterm.js.
Proxmox VE Administration Guide 80 / 464
On the right side the content switches depending on the selected option.
On the left side. All available options are listed one below the other.
• Task History: here all previous tasks from the selected guest will be shown.
• Backup: shows the available backups from the selected guest and also create a backupset.
• Replication: shows the replication jobs for the selected guest and allows to create new jobs.
4.4.4 Storage
In this view we have a two partition split-view. On the left side we have the storage options and on the right
side the content of the selected option will be shown.
• Summary: shows important information about storages like Usage, Type, Content, Active and Enabled.
4.4.5 Pools
In this view we have a two partition split view. On the left side we have the logical pool options and on the
right side the content of the selected option will be shown.
• Members: Here all members of this pool will listed and can be managed.
Chapter 5
Cluster Manager
The Proxmox VE cluster manager pvecm is a tool to create a group of physical servers. Such a group is
called a cluster. We use the Corosync Cluster Engine for reliable group communication, and such clusters
can consist of up to 32 physical nodes (probably more, dependent on network latency).
pvecm can be used to create a new cluster, join nodes to a cluster, leave the cluster, get status informa-
tion and do various other cluster related tasks. The Proxmox Cluster File System (“pmxcfs”) is used to
transparently distribute the cluster configuration to all cluster nodes.
Grouping nodes into a cluster has the following advantages:
• pmxcfs: database-driven file system for storing configuration files, replicated in real-time on all nodes
using corosync.
• Fast deployment
5.1 Requirements
• All nodes must be able to connect to each other via UDP ports 5404 and 5405 for corosync to work.
• If you are interested in High Availability, you need to have at least three nodes for reliable quorum. All
nodes should have the same version.
• We recommend a dedicated NIC for the cluster traffic, especially if you use shared storage.