System Deployment¶
The DAOS deployment workflow requires to start the DAOS server instances early on to enable administrators to perform remote operations in parallel across multiple storage nodes via the dmg management utility. Security is guaranteed via the use of certificates. The first type of commands run after installation include network and storage hardware provisioning and would typically be run from a login node.
After daos_server
instances have been started on each storage node for the
first time, daos_server scm prepare
will set PMem storage into the necessary
state for use with DAOS when run on each host.
Then dmg storage format
formats persistent storage devices (specified in the
server configuration file) on the storage nodes and writes necessary metadata
before starting DAOS I/O engine processes that will operate across the fabric.
To sum up, the typical workflow of a DAOS system deployment consists of the following steps:
-
Configure and start the DAOS server.
-
Provision Hardware on all the storage nodes via the dmg utility.
-
Format the DAOS system
-
Set up and start the agent on the client nodes
-
Validate that the DAOS system is operational
Note
Starting the DAOS server instances can be performed automatically on boot if start-up scripts are registered with systemd.
The following subsections will cover each step in more detail.
DAOS Server Setup¶
First of all, the DAOS server should be started to allow remote administration command to be executed via the dmg tool. This section describes the minimal DAOS server configuration and how to start it on all the storage nodes.
Example RPM Deployment Workflow¶
A recommended workflow to get up and running is as follows:
-
Install DAOS Server RPMs -
daos_server
systemd services will start in listening mode which means DAOS I/O engine processes will not be started as the server config file (default location at/etc/daos/daos_server.yml
) has not yet been populated. -
Run
dmg config generate -l <hostset> -a <access_points>
across the entire hostset (all the storage servers that are now running thedaos_server
service after RPM install). The command will only generate a config if hardware setups on all the hosts are similar and have been given sensible NUMA mappings. Adjust the hostset until you have a set with homogeneous hardware configurations. -
Once a recommended config file can be generated, copy it to the server config file default location (
/etc/daos/daos_server.yml
) on each DAOS Server host and restart alldaos_server
services. An example command to restart the services isclush -w machines-[118-121,130-133] "sudo systemctl restart daos_server"
. The services should prompt for format on restart and after format is triggered fromdmg
, the DAOS I/O engine processes should start.
Server Configuration File¶
The daos_server
configuration file is parsed when starting the daos_server
process.
The configuration file location can be specified on the command line
(daos_server -h
for usage) otherwise it will be read from the default location
(/etc/daos/daos_server.yml
).
Parameter descriptions are specified in
daos_server.yml
and example configuration files in the
examples
directory.
Any option supplied to daos_server
as a command line option or flag will
take precedence over equivalent configuration file parameter.
For convenience, active parsed configuration values are written to a temporary file for reference, and the location will be written to the log.
Configuration Options¶
The example configuration file lists the default empty configuration, listing all the options (living documentation of the config file). Live examples are available at https://github.com/daos-stack/daos/tree/master/utils/config/examples
The location of this configuration file is determined by first checking
for the path specified through the -o option of the daos_server
command
line, if unspecified then /etc/daos/daos_server.yml
is used.
Refer to the example configuration file
daos_server.yml
for latest information and examples.
MD-on-SSD Configuration¶
To enable MD-on-SSD, the Control-Plane-Metadata ('control_metadata') global section of the
configuration file
daos_server.yml
needs to specify a persistent location to store control-plane specific metadata (which would be
stored on PMem in non MD-on-SSD mode). Either set 'control_metadata:path' to an existing (mounted)
local filesystem path or set 'control_metadata:device' to a storage partition which can be mounted
and formatted by the control-plane during storage format. In the latter case when specifying a
device the path parameter value will be used as the mountpoint path.
The MD-on-SSD code path will only be used if it is explicitly enabled by specifying the new 'bdev_role' property for the NVMe storage tier(s) in the 'daos_server.yml' file. There are three types of 'bdev_role': wal, meta, and data. Each role must be assigned to exactly one NVMe tier. Depending on the number of NVMe SSDs per DAOS engine there may be one, two or three NVMe tiers with different 'bdev_role' assignments.
For a complete server configuration file example enabling MD-on-SSD, see
daos_server_mdonssd.yml
.
Below are four different 'daos_server.yml' storage configuration snippets that represent scenarios for a DAOS engine with four NVMe SSDs and MD-on-SSD enabled.
- One NVMe tier where four SSDs are each assigned wal, meta and data roles:
storage:
-
class: ram
scm_mount: /mnt/dram1
-
class: nvme
bdev_roles:
- wal
- meta
- data
bdev_list:
- "0000:e3:00.0"
- "0000:e4:00.0"
- "0000:e5:00.0"
- "0000:e6:00.0"
This example shows the typical use case for a DAOS server with a small number of NVMe SSDs. With only four or five NVMe SSDs per engine, it is natural to assign all three roles to all NVMe SSDs configured as a single NVMe tier.
- Two NVMe tiers, one SSD assigned wal role (tier-1) and three SSDs assigned both meta and data roles (tier-2):
storage:
-
class: ram
scm_mount: /mnt/dram1
-
class: nvme
bdev_roles:
- wal
bdev_list:
- "0000:e3:00.0"
-
class: nvme
bdev_roles:
- meta
- data
bdev_list:
- "0000:e4:00.0"
- "0000:e5:00.0"
- "0000:e6:00.0"
This example shows where one NVMe SSD is dedicated for the wal, while the remaining three NVMe SSDs are assigned to hold the VOS checkpoints (meta) and the user data. Using two NVMe tiers makes it possible to use a higher endurance and higher performance SSD for the wal tier. It should be noted that the performance of a single high-performance SSD may still be lower than the aggregate performance of multiple lower-performance SSDs in the previous scenario.
- Two NVMe tiers, one SSD assigned both wal and meta roles (tier-1) and three SSDs assigned both meta and data roles (tier-2):
storage:
-
class: ram
scm_mount: /mnt/dram1
-
class: nvme
bdev_roles:
- wal
- meta
bdev_list:
- "0000:e3:00.0"
-
class: nvme
bdev_roles:
- data
bdev_list:
- "0000:e4:00.0"
- "0000:e5:00.0"
- "0000:e6:00.0"
This example uses two NVMe tiers but co-locates the wal and meta blobs on the same tier. This may be a better choice than co-locating meta and data if the endurance of the data NVMe SSDs is too low for the relatively frequent VOS checkpointing. It should be noted that performance may be impacted by reducing the number of NVMe SSDs available for VOS checkpointing from three to one.
The other option to use two NVMe tiers would be to co-locate wal and data and dedicate the other tier for meta but this configuration is invalid. Sharing wal and data roles on the same tier is not allowed.
- Three NVMe tiers, one for each role (each device will have a single distinct role), one SSD assigned wal role (tier-1), one SSD assigned meta role (tier-2) and two SSDs assigned data role (tier-3):
storage:
-
class: ram
scm_mount: /mnt/dram1
-
class: nvme
bdev_roles:
- wal
bdev_list:
- "0000:e3:00.0"
-
class: nvme
bdev_roles:
- meta
bdev_list:
- "0000:e4:00.0"
-
class: nvme
bdev_roles:
- data
bdev_list:
- "0000:e5:00.0"
- "0000:e6:00.0"
Using three NVMe tiers (one per role) is not reasonable when there are only four NVMe SSDs per engine but maybe practical with a larger number of SSDs and so illustrated here for completeness.
Auto Generate Configuration File¶
DAOS can attempt to produce a server configuration file that makes optimal use of hardware on a
given set of hosts either through the dmg
or daos_server
tools.
To generate an MD-on-SSD configurations set both '--control-metadata-path' and '--use-tmpfs-scm' options as detailed below. Note that due to the number of variables considered when generating a configuration automatically the result may not be the most optimal in all situations.
Generating Configuration File Using daos_server Tool¶
To generate a configuration file for a single storage server, run the daos_server config generate
command locally. In this case, the daos_server
service should not be running on the local host.
daos_server scm prepare
should be run prior to daos_server config generate
if the intent is to
use PMem for SCM. This will allow the relevant storage to be available for detection when
attempting to generate the configuration files.
$ daos_server config generate --help
Usage:
daos_server [OPTIONS] config generate [generate-OPTIONS]
Application Options:
--allow-proxy Allow proxy configuration via environment
-o, --config= Server config file path
-b, --debug Enable debug output
-j, --json Enable JSON output
-J, --json-logging Enable JSON-formatted log output
--syslog Enable logging to syslog
Help Options:
-h, --help Show this help message
[generate command options]
-l, --helper-log-file= Log file location for debug from daos_server_helper binary
-a, --access-points= Comma separated list of access point addresses
<ipv4addr/hostname> (default: localhost)
-e, --num-engines= Set the number of DAOS Engine sections to be populated in the
config file output. If unset then the value will be set to the
number of NUMA nodes on storage hosts in the DAOS system.
-s, --scm-only Create a SCM-only config without NVMe SSDs.
-c, --net-class=[ethernet|infiniband] Set the network device class to be used (default: infiniband)
-p, --net-provider= Set the network fabric provider to be used
-t, --use-tmpfs-scm Use tmpfs for scm rather than PMem
-m, --control-metadata-path= External storage path to store control metadata. Set this to
a persistent location and specify --use-tmpfs-scm to create an
MD-on-SSD config
-f, --fabric-ports= Allow custom fabric interface ports to be specified for each engine
config section. Comma separated port numbers, one per engine
--skip-prep Skip preparation of devices during scan.
Note the --helper-log-file
which can be used to provide a log file path to output debug level
logging from the privileged server helper binary. This can be used for troubleshooting in addition to
the main application debug flag.
Generating Configuration File Using dmg Tool¶
To generate a configuration file for a group of storage servers with homogeneous hardware installed,
dmg config generate
command can be called which will operate over remote addresses specified in
the --host-list
application option (the hostlist can also be specified in the dmg client config).
$ dmg config generate --help
Usage:
dmg [OPTIONS] config generate [generate-OPTIONS]
Application Options:
--allow-proxy Allow proxy configuration via environment
-i, --insecure Have dmg attempt to connect without certificates
-d, --debug Enable debug output
--log-file= Log command output to the specified file
-j, --json Enable JSON output
-J, --json-logging Enable JSON-formatted log output
-o, --config-path= Client config file path
Help Options:
-h, --help Show this help message
[generate command options]
-l, --host-list= A comma separated list of addresses <ipv4addr/hostname> to connect to
-a, --access-points= Comma separated list of access point addresses <ipv4addr/hostname>
to host management service (default: localhost)
-e, --num-engines= Set the number of DAOS Engine sections to be populated in the
config file output. If unset then the value will be set to the
number of NUMA nodes on storage hosts in the DAOS system.
-s, --scm-only Create a SCM-only config without NVMe SSDs.
-c, --net-class=[ethernet|infiniband] Set the network device class to be used (default: infiniband)
-p, --net-provider= Set the network fabric provider to be used
-t, --use-tmpfs-scm Use tmpfs for scm rather than PMem
-m, --control-metadata-path= External storage path to store control metadata. Set this to
a persistent location and specify --use-tmpfs-scm to create an
MD-on-SSD config
-f, --fabric-ports= Allow custom fabric interface ports to be specified for each engine
config section. Comma separated port numbers, one per engine
The daos_server
service must be running on the remote storage servers and as such a minimal
server config file must already exist.
An example of a minimal config file is as follows which will enable basic validation to pass and the
daos_server
to start (see the Server startup section for help on starting the
service):
$ cat /etc/daos/daos_server.yml
provider: ofi+tcp
engines:
- provider: ofi+tcp
fabric_iface: ib0
fabric_iface_port: 31416
storage:
- class: ram
scm_mount: /mnt/daos0
scm_size: 16
Config Generate Command Operation¶
The options that can be supplied to the config generate command are as follows:
-
--access-points
specifies the access points (identified storage servers that will host the management service for the DAOS system across the cluster). -
--num-engines
specifies the number of engine sections to populate in the config file output. If not set explicitly on the commandline, default is the number of NUMA nodes detected on the host. Each generated engine section will specify a SCM storage tier (PMem or tmpfs) in addition to one or more NVMe storage tiers. All hardware components specified in an engine config section should be bound to the same NUMA node (PMem bdev, SSDs and host fabric interface). -
--scm-only
requests that a config without NVMe should be generated. This flag will override the command's normal behavior and should be used only in circumstances where NVMe SSDs are unavailable or not balanced across NUMA nodes and multiple engines are required per host. Note that DAOS performance will be suboptimal without NVMe SSDs. -
--net-class
selects a specific network device class, options areethernet
orinfiniband
. If not set explicitly on the commandline, default isinfiniband
. This option will alter the command's behavior and should only be used when normal command operation is not sufficient. -
--net-provider
selects a specific network provider , this can be set to any provider string supported on the hosts e.g. ofi+tcp. This option will alter the command's behavior and should only be used when normal command operation is not sufficient. Note that specifying a provider will prevent the command from selecting the best available. -
--use-tmpfs-scm
will produce a config specifying RAM-disk (tmpfs) devices in the first (SCM) storage tier. The RAM-disk sizes will be calculated based on the host's total memory (as reported by/proc/meminfo
). -
--control-metadata-path
specifies a persistent location to store control-plane metadata which allows MD-on-SSD DAOS deployments to survive without data loss over 'daos_server' restarts. If this option is set then a MD-on-SSD config will be generated. -
--fabric-ports
enables custom port numbers to be assigned to each engine's fabric settings. Comma separated list must contain enough numbers to cover all engines generated in config.
The text generated by the command and output to stdout can be copied and used as the server config
file on relevant hosts (normally by copying to /etc/daos/daos_server.yml
and (re)starting service).
Config Generate Command Troubleshooting¶
The config generate command may fail to generate output in the following cases:
-
When running with the
dmg
tool, if installed hardware device count or NUMA mappings differ on any of the hosts in the hostlist. The output ofdaos_server (scm|nvme|network) scan
can be used to detect hardware differences between hosts, examples of differences that might prevent a config from being generated are NVMe SSD count, PCI address distribution or device NUMA affinities. -
NUMA node count can't be detected on the hosts or differs on any host in the host list.
-
NUMA node count doesn't meet the
num-engines
requirement when--use-tmpfs-scm
is specified. -
PMem device count or NUMA affinity doesn't meet the
num-engines
requirement. -
NVMe device NUMA affinity imbalanced (or all bound to one socket).
-
Network device count or NUMA affinity doesn't match the
num-engines
requirement. Limitations regarding network device class and provider support should also be taken into account.
Config Generate Example Usage¶
The following example executes the daos_server config generate
local command with commandline
options to select a specific provider and use RAM-disk for SCM. The storage server has balanced
NUMA Affinity for hardware components and therefore successfully generates a config with one
fabric interface per socket/NUMA (one NUMA node per socket on the target host) and 8 NVMe SSDs
per engine. The command redirects stderr to /dev/null and stdout to a temporary file. The
installation is from a source build.
[user@wolf-226 daos]$ install/bin/daos_server config generate -p ofi+tcp --use-tmpfs-scm 2>/dev/null | tee ~/configs/tmp.yml
port: 10001
transport_config:
allow_insecure: false
client_cert_dir: /etc/daos/certs/clients
ca_cert: /etc/daos/certs/daosCA.crt
cert: /etc/daos/certs/server.crt
key: /etc/daos/certs/server.key
engines:
- targets: 16
nr_xs_helpers: 4
first_core: 0
log_file: /tmp/daos_engine.0.log
storage:
- class: ram
scm_mount: /mnt/daos0
scm_size: 85
- class: nvme
bdev_list:
- 0000:2b:00.0
- 0000:2c:00.0
- 0000:2d:00.0
- 0000:2e:00.0
- 0000:63:00.0
- 0000:64:00.0
- 0000:65:00.0
- 0000:66:00.0
provider: ofi+tcp
fabric_iface: ib0
fabric_iface_port: 31416
pinned_numa_node: 0
- targets: 16
nr_xs_helpers: 4
first_core: 0
log_file: /tmp/daos_engine.1.log
storage:
- class: ram
scm_mount: /mnt/daos1
scm_size: 85
- class: nvme
bdev_list:
- 0000:a2:00.0
- 0000:a3:00.0
- 0000:a4:00.0
- 0000:a5:00.0
- 0000:de:00.0
- 0000:df:00.0
- 0000:e0:00.0
- 0000:e1:00.0
provider: ofi+tcp
fabric_iface: ib1
fabric_iface_port: 32416
pinned_numa_node: 1
disable_vfio: false
disable_vmd: false
enable_hotplug: false
nr_hugepages: 16384
disable_hugepages: false
control_log_mask: INFO
control_log_file: /tmp/daos_server.log
core_dump_filter: 19
name: daos_server
socket_dir: /var/run/daos_server
provider: ofi+tcp
access_points:
- localhost:10001
fault_cb: ""
hyperthreads: false
Now we start the daos_server
service from the generated config which loads successfully
and runs until the point where a storage format is required, as expected.
[user@wolf-226 daos]$ install/bin/daos_server start -i -o ~/configs/tmp.yml
DAOS Server config loaded from /home/user/configs/tmp.yml
install/bin/daos_server logging to file /tmp/daos_server.log
NOTICE: Configuration includes only one access point. This provides no redundancy in the event of an access point failure.
DAOS Control Server v2.3.101 (pid 1211553) listening on 127.0.0.1:10001
Checking DAOS I/O Engine instance 0 storage ...
Checking DAOS I/O Engine instance 1 storage ...
SCM format required on instance 0
SCM format required on instance 1
The daos_server start --auto-format
option has now been added to allow automatic formatting on
startup in the event that externally triggered format is not desirable. When using this option
the storage format will occur on daos_server
startup without any user intervention assuming the
supplied config file is valid.
To manually format the storage and start the engine processes, we run the following on a separate terminal window and verify that engine processes (ranks) have registered with the system. Note the subsequent system query command may not show ranks started immediately after the storage format command returns so it is recommended to leave a short delay (~5s) before invoking.
[user@wolf-226 daos]$ install/bin/dmg storage format -i
Format Summary:
Hosts SCM Devices NVMe Devices
----- ----------- ------------
localhost 2 16
[user@wolf-226 daos]$ install/bin/dmg system query -i
Rank State
---- -----
[0-1] Joined
Engine I/O processes will be launched by the daos_server
service following storage format.
Instance 0: starting format of nvme block devices 0000:2b:00.0,0000:2c:00.0,0000:2d:00.0,0000:2e:00.0,0000:63:00.0,0000:64:00.0,0000:65:00.0,0000:66:00.0
Instance 0: finished format of nvme block devices 0000:2b:00.0,0000:2c:00.0,0000:2d:00.0,0000:2e:00.0,0000:63:00.0,0000:64:00.0,0000:65:00.0,0000:66:00.0
Format of NVMe storage for DAOS I/O Engine instance 0: 6.15314273s
Writing NVMe config file for engine instance 0 to "/mnt/daos0/daos_nvme.conf"
Instance 1: starting format of nvme block devices 0000:a2:00.0,0000:a3:00.0,0000:a4:00.0,0000:a5:00.0,0000:de:00.0,0000:df:00.0,0000:e0:00.0,0000:e1:00.0
Instance 1: finished format of nvme block devices 0000:a2:00.0,0000:a3:00.0,0000:a4:00.0,0000:a5:00.0,0000:de:00.0,0000:df:00.0,0000:e0:00.0,0000:e1:00.0
Format of NVMe storage for DAOS I/O Engine instance 1: 6.15925073s
Writing NVMe config file for engine instance 1 to "/mnt/daos1/daos_nvme.conf"
DAOS I/O Engine instance 0 storage ready
DAOS I/O Engine instance 1 storage ready
SCM @ /mnt/daos1: 91 GB Total/91 GB Avail
Starting I/O Engine instance 1: /home/user/projects/daos/install/bin/daos_engine
daos_engine:1 Using NUMA core allocation algorithm
SCM @ /mnt/daos0: 91 GB Total/91 GB Avail
Starting I/O Engine instance 0: /home/user/projects/daos/install/bin/daos_engine
daos_engine:0 Using NUMA core allocation algorithm
MS leader running on wolf-226.wolf.hpdd.intel.com
daos_engine:1 DAOS I/O Engine (v2.3.101) process 1215202 started on rank 1 with 16 target, 4 helper XS, firstcore 0, host wolf-226.wolf.hpdd.intel.com.
Using NUMA node: 1
daos_engine:0 DAOS I/O Engine (v2.3.101) process 1215209 started on rank 0 with 16 target, 4 helper XS, firstcore 0, host wolf-226.wolf.hpdd.intel.com.
Using NUMA node: 0
For reference, the hardware scan results for the target storage server are included below.
[user@wolf-226 daos]$ install/bin/daos_server nvme scan
Scan locally-attached NVMe storage...
NVMe PCI Model FW Revision Socket ID Capacity
-------- ----- ----------- --------- --------
0000:2b:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:2c:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:2d:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:2e:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:63:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:64:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:65:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:66:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 0 3.8 TB
0000:a2:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:a3:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:a4:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:a5:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:de:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:df:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:e0:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
0000:e1:00.0 MZXLR3T8HBLS-000H3 MPK7525Q 1 3.8 TB
[user@wolf-226 daos]$ install/bin/daos_server network scan
---------
localhost
---------
-------------
NUMA Socket 0
-------------
Provider Interfaces
-------- ----------
ofi+tcp;ofi_rxm eth0, ib0
ucx+all eth0, ib0, ib0
ucx+dc ib0
ucx+rc_x ib0
ucx+tcp eth0, ib0
ofi+sockets eth0, ib0
ucx+dc_x ib0
ucx+rc_v ib0
ucx+ud ib0
ofi+verbs;ofi_rxm ib0
ucx+rc ib0
ucx+ud_v ib0
ucx+ud_x ib0
udp;ofi_rxd eth0, ib0
udp eth0, ib0
ofi+tcp eth0, ib0
ofi+verbs ib0
-------------
NUMA Socket 1
-------------
Provider Interfaces
-------- ----------
ofi+tcp ib1
ofi+verbs;ofi_rxm ib1
ucx+dc_x ib1
ucx+rc_x ib1
ucx+tcp ib1
ucx+ud ib1
ofi+sockets ib1
udp;ofi_rxd ib1
udp ib1
ucx+dc ib1
ucx+all ib1, ib1
ofi+verbs ib1
ucx+rc ib1
ucx+rc_v ib1
ucx+ud_v ib1
ucx+ud_x ib1
ofi+tcp;ofi_rxm ib1
Certificate Configuration¶
The DAOS security framework relies on certificates to authenticate
components and administrators in addition to encrypting DAOS control plane
communications. A set of certificates for a given DAOS system may be
generated by running the gen_certificates.sh
script provided with the DAOS
software if there is not an existing TLS certificate infrastructure. The
gen_certificates.sh
script uses the openssl
tool to generate all of the
necessary files. We highly recommend using OpenSSL Version 1.1.1h or higher as
keys and certificates generated with earlier versions are vulnerable to attack.
When DAOS is installed from RPMs, this script is provided in the base daos
RPM, and
may be invoked in the directory to which the certificates will be written. As part
of the generation process, a new local Certificate Authority is created to handle
certificate signing, and three role certificates are created:
# /usr/lib64/daos/certgen/gen_certificates.sh
Generating Private CA Root Certificate
Private CA Root Certificate created in ./daosCA
...
Generating Server Certificate
Required Server Certificate Files:
./daosCA/certs/daosCA.crt
./daosCA/certs/server.key
./daosCA/certs/server.crt
...
Generating Agent Certificate
Required Agent Certificate Files:
./daosCA/certs/daosCA.crt
./daosCA/certs/agent.key
./daosCA/certs/agent.crt
...
Generating Admin Certificate
Required Admin Certificate Files:
./daosCA/certs/daosCA.crt
./daosCA/certs/admin.key
./daosCA/certs/admin.crt
The files generated under ./daosCA should be protected from unauthorized access and preserved for future use.
The generated keys and certificates must then be securely distributed to all nodes participating in the DAOS system (servers, clients, and admin nodes). Permissions for these files should be set to prevent unauthorized access to the keys and certificates.
Client nodes require:
- CA root cert
- Agent cert
- Agent key
Administrative nodes require:
- CA root cert
- Admin cert
- Admin key
Server nodes require:
- CA root cert
- Server cert
- Server key
- All valid agent certs in the DAOS system (in the client cert directory, see config file below)
After the certificates have been securely distributed, the DAOS configuration files must be
updated in order to enable authentication and secure communications. These examples assume
that the configuration and certificate files have been installed under /etc/daos
:
# /etc/daos/daos_server.yml (servers)
transport_config:
# Location where daos_server will look for Client certificates
client_cert_dir: /etc/daos/certs/clients
# Custom CA Root certificate for generated certs
ca_cert: /etc/daos/certs/daosCA.crt
# Server certificate for use in TLS handshakes
cert: /etc/daos/certs/server.crt
# Key portion of Server Certificate
key: /etc/daos/certs/server.key
# /etc/daos/daos_agent.yml (clients)
transport_config:
# Custom CA Root certificate for generated certs
ca_cert: /etc/daos/certs/daosCA.crt
# Agent certificate for use in TLS handshakes
cert: /etc/daos/certs/agent.crt
# Key portion of Agent Certificate
key: /etc/daos/certs/agent.key
# /etc/daos/daos_control.yml (dmg/admin)
transport_config:
# Custom CA Root certificate for generated certs
ca_cert: /etc/daos/certs/daosCA.crt
# Admin certificate for use in TLS handshakes
cert: /etc/daos/certs/admin.crt
# Key portion of Admin Certificate
key: /etc/daos/certs/admin.key
Server Startup¶
The DAOS Server is started as a systemd service. The DAOS Server
unit file is installed in the correct location when installing from RPMs.
The DAOS Server will be run as daos-server
user which will be created
during RPM install.
If you wish to use systemd with a development build, you must copy the service
file from utils/systemd to /usr/lib/systemd/system
. Once the file is copied,
modify the ExecStart line to point to your daos_server
binary.
After modifying ExecStart, run the following command:
$ sudo systemctl daemon-reload
Once the service file is installed you can start daos_server
with the following commands:
$ systemctl enable daos_server.service
$ systemctl start daos_server.service
To check the component status use:
$ systemctl status daos_server.service
If DAOS Server failed to start, check the logs with:
$ journalctl --unit daos_server.service
After RPM install, daos_server
service starts automatically running as user
"daos". The server config is read from /etc/daos/daos_server.yml
and
certificates are read from /etc/daos/certs
.
With no other admin intervention other than the loading of certificates,
daos_server
will enter a listening state enabling discovery of storage and
network hardware through the dmg
tool without any I/O engines specified in the
configuration file. After device discovery and provisioning, an updated
configuration file with a populated per-engine section can be stored in
/etc/daos/daos_server.yml
, and after reestarting the daos_server
service
it is then ready for the storage to be formatted.
DAOS Server Remote Access¶
Remote tasking of the DAOS system and individual DAOS Server processes can be
performed via the dmg
utility.
To set the addresses of which DAOS Servers to task, provide either:
-
-l <hostlist>
on the commandline when invoking, or -
hostlist: <hostlist>
in the control configuration filedaos_control.yml
Where <hostlist>
represents a slurm-style hostlist string e.g.
foo-1[28-63],bar[256-511]
.
The first entry in the hostlist (after alphabetic then numeric sorting) will be
assumed to be the access point as set in the server configuration file.
Local configuration files stored in the user directory will be used in
preference to the default location e.g. ~/.daos_control.yml
.
Hardware Provisioning¶
Once the DAOS server started, the storage and network can be configured on the storage nodes via the dmg utility.
Note
daos_server (nvme|scm)
commands are config aware meaning they will read parameters from the
server configuration file unless the --ignore-config
flag is supplied.
SCM Preparation¶
This section addresses how to verify that PMem (Intel(R) Optane(TM) persistent memory) modules are correctly installed on the storage nodes and how to configure them in interleaved mode to be used by DAOS. Instructions for other types of SCM may be covered in the future.
Provisioning the SCM occurs by configuring PMem modules in interleaved memory regions (interleaved mode) in groups of modules local to a specific socket (NUMA), and resultant PMem namespaces are defined by a device identifier (e.g., /dev/pmem0).
PMem preparation is required once per DAOS installation.
This step requires a reboot to enable PMem resource allocation changes to be read by BIOS.
PMem preparation can be performed with daos_server scm prepare
.
The first time the command is run, the SCM interleaved regions will be created as resource allocations on any available PMem modules (one region per socket). The regions are activated after BIOS reads the new resource allocations. Upon completion, the scm prepare command will prompt the admin to reboot the storage node(s) in order for the BIOS to activate the new storage allocations. The scm prepare command does not initiate the reboot itself.
After running the command a reboot will be required, the command will then need to be run for a second time to expose the namespace device to be used by DAOS.
Example usage:
-
clush -w wolf-[118-121,130-133] daos_server scm prepare
after running, the user should be prompted for a reboot. -
clush -w wolf-[118-121,130-133] reboot
-
clush -w wolf-[118-121,130-133] daos_server scm prepare
after running, PMem namespaces (/dev/pmemX block devices created on the new SCM regions) should be available on each of the hosts.
On the second run, one namespace per region is created, and each namespace may take up to a few minutes to create. Details of the PMem namespaces will be displayed in JSON format on command completion.
Upon successful creation of the PMem namespaces, the Intel(R) Optane(TM) persistent memory is configured and one can move on to the next step.
If required, the PMem namespaces can be destroyed with the command
daos_server scm reset
.
All namespaces are disabled and destroyed. The SCM regions are removed by resetting modules into "MemoryMode" through resource allocations.
Warning
Undefined behavior may result if the namespaces/pmem kernel block devices are mounted when running the reset command (as per the printed warning).
A subsequent reboot is required for BIOS to read the new resource allocations.
Multiple PMem namespaces per socket¶
By default the daos_server scm prepare
command will create one PMem namespace on each PMem
region.
A single PMem AppDirect region will be created for each NUMA node (typically one NUMA node per CPU
socket) in interleaved mode (which indicates that all PMem modules attached to a particular socket
will be used in a single set/region).
Therefore by default on a dual-socket platform, two regions and two namespaces will be created.
Multiple PMem namespaces can be created on a single region (one per socket) by specifying a value
of 1-8 in (-S|--scm-ns-per-socket)
commandline option for daos_server scm prepare
subcommand.
Example usage:
$ daos_server scm prepare -f -S 4
Prepare locally-attached SCM...
Memory allocation goals for PMem will be changed and namespaces modified, this may be a destructive
operation. Please ensure namespaces are unmounted and locally attached PMem modules are not in use.
Please be patient as it may take several minutes and subsequent reboot maybe required.
SCM Namespace Socket ID Capacity
------------- --------- --------
pmem0 0 796 GB
pmem0.1 0 796 GB
pmem0.2 0 796 GB
pmem0.3 0 796 GB
pmem1 1 796 GB
pmem1.1 1 796 GB
pmem1.2 1 796 GB
pmem1.3 1 796 GB
Note
This feature is in a beta phase and not supported in production deployments.
Storage Discovery and Selection¶
This section covers how to manually detect and select storage devices to be used by DAOS. The server configuration file gives an administrator the ability to control storage selection.
Note
daos_server (nvme|scm)
commands are config aware meaning they will read parameters from the
server configuration file unless the --ignore-config
flag is supplied.
Discovery¶
DAOS tools will discover NVMe SSDs and Persistent Memory Modules using the storage scan commands.
dmg storage scan
can be run to query remote running daos_server
processes over the management
network.
daos_server (nvme|scm) scan
can be used to query storage on the local host directly.
Note
'daos_server' commands will refuse to run if a process with the same name exists (e.g. as a systemd service under the 'daos_server' userid).
NVMe SSDs no longer need to be made accessible first by running daos_server nvme prepare
,
daos_server nvme scan
will take the necessary steps to prepare the devices unless --skip-prep
flag is supplied.
The default way for DAOS to access NVMe storage is through SPDK via the VFIO user-space driver.
To use an alternative driver with SPDK, set --disable-vfio
in the nvme prepare command to
fallback to using UIO user-space driver with SPDK instead.
Note
If UIO user-space driver is used instead of VFIO, 'daos_server' needs to be run as root.
The output will be equivalent running dmg storage scan --verbose
remotely.
bash-4.2$ dmg storage scan
Hosts SCM Total NVMe Total
----- --------- ----------
wolf-[71-72] 6.4 TB (2 namespaces) 3.1 TB (3 controllers)
bash-4.2$ dmg storage scan --verbose
------------
wolf-[71-72]
------------
SCM Namespace Socket ID Capacity
------------- --------- --------
pmem0 0 3.2 TB
pmem1 1 3.2 TB
NVMe PCI Model FW Revision Socket ID Capacity
-------- ----- ----------- --------- --------
0000:81:00.0 INTEL SSDPED1K750GA E2010325 1 750 GB
0000:87:00.0 INTEL SSDPEDMD016T4 8DV10171 1 1.6 TB
0000:da:00.0 INTEL SSDPED1K750GA E2010325 1 750 GB
The NVMe PCI field above is what should be used in the server configuration file to identified NVMe SSDs.
Devices with the same NUMA node/socket should be used in the same per-engine section of the server configuration file for best performance.
For further info on dmg storage command usage run dmg storage --help
.
To release the NVMe drives from the user-space drivers and bind them back to the kernel "nvme"
driver so they can be used by the OS (and reappear as /dev/nvme* block devices), run
daos_server nvme reset
with any relevant options (see command help for details). The reset
command will also release any hugepages used by no-longer-active SPDK processes.
daos_server nvme scan
will perform a subsequent reset implicitly so manual reset is not required
but a stopped daos_server
may not reset device bindings and hugepage resources and may require a
manual reset to do so.
Warning
Due to SPDK issue 2926, if VMD is enabled and
PCI_ALLOWED list is set to a subset of available VMD controllers (as specified in the server
config file) then the backing devices of the unselected VMD controllers will be bound to no
driver and therefore inaccessible from both OS and SPDK. Workaround is to run
daos_server nvme scan --ignore-config
to reset driver bindings for all VMD controllers.
Health¶
SSD health state can be verified via dmg storage scan --nvme-health
:
bash-4.2$ dmg storage scan --nvme-health
-------
wolf-71
-------
PCI:0000:81:00.0 Model:INTEL SSDPED1K750GA FW:E2010325 Socket:1 Capacity:750 GB
Health Stats:
Timestamp:2021-09-13T11:12:34.000+00:00
Temperature:318K(44.85C)
Controller Busy Time:0s
Power Cycles:15
Power On Duration:10402h0m0s
Unsafe Shutdowns:13
Error Count:0
Media Errors:0
Read Errors:0
Write Errors:0
Unmap Errors:0
Checksum Errors:0
Error Log Entries:0
Critical Warnings:
Temperature: OK
Available Spare: OK
Device Reliability: OK
Read Only: OK
Volatile Memory Backup: OK
Intel Vendor SMART Attributes:
Program Fail Count:
Normalized:100%
Raw:0
Erase Fail Count:
Normalized:100%
Raw:0
Wear Leveling Count:
Normalized:100%
Min:0
Max:9
Avg:3
End-to-End Error Detection Count:0
CRC Error Count:0
Timed Workload, Media Wear:65535
Timed Workload, Host Read/Write Ratio:65535
Timed Workload, Timer:65535
Thermal Throttle Status:0%
Thermal Throttle Event Count:0
Retry Buffer Overflow Counter:0
PLL Lock Loss Count:0
NAND Bytes Written:222897
Host Bytes Written:71
PCI:0000:da:00.0 Model:INTEL SSDPED1K750GA FW:E2010325 Socket:1 Capacity:750 GB
Health Stats:
Temperature:320K(46.85C)
Controller Busy Time:0s
Power Cycles:15
Power On Duration:10402h0m0s
Unsafe Shutdowns:13
Error Count:0
Media Errors:0
Read Errors:0
Write Errors:0
Unmap Errors:0
Checksum Errors:0
Error Log Entries:0
Critical Warnings:
Temperature: OK
Available Spare: OK
Device Reliability: OK
Read Only: OK
Volatile Memory Backup: OK
Intel Vendor SMART Attributes:
Program Fail Count:
Normalized:100%
Raw:0
Erase Fail Count:
Normalized:100%
Raw:0
Wear Leveling Count:
Normalized:100%
Min:0
Max:9
Avg:3
End-to-End Error Detection Count:0
CRC Error Count:0
Timed Workload, Media Wear:65535
Timed Workload, Host Read/Write Ratio:65535
Timed Workload, Timer:65535
Thermal Throttle Status:0%
Thermal Throttle Event Count:0
Retry Buffer Overflow Counter:0
PLL Lock Loss Count:0
NAND Bytes Written:222897
Host Bytes Written:71
-------
wolf-72
-------
PCI:0000:81:00.0 Model:INTEL SSDPED1K750GA FW:E2010435 Socket:1 Capacity:750 GB
Health Stats:
Temperature:316K(42.85C)
Controller Busy Time:8m0s
Power Cycles:23
Power On Duration:10399h0m0s
Unsafe Shutdowns:18
Error Count:0
Media Errors:0
Read Errors:0
Write Errors:0
Unmap Errors:0
Checksum Errors:0
Error Log Entries:0
Critical Warnings:
Temperature: OK
Available Spare: OK
Device Reliability: OK
Read Only: OK
Volatile Memory Backup: OK
Intel Vendor SMART Attributes:
Program Fail Count:
Normalized:100%
Raw:0
Erase Fail Count:
Normalized:100%
Raw:0
Wear Leveling Count:
Normalized:100%
Min:0
Max:9
Avg:3
End-to-End Error Detection Count:0
CRC Error Count:0
Timed Workload, Media Wear:65535
Timed Workload, Host Read/Write Ratio:65535
Timed Workload, Timer:65535
Thermal Throttle Status:0%
Thermal Throttle Event Count:0
Retry Buffer Overflow Counter:0
PLL Lock Loss Count:0
NAND Bytes Written:222897
Host Bytes Written:71
PCI:0000:da:00.0 Model:INTEL SSDPED1K750GA FW:E2010435 Socket:1 Capacity:750 GB
Health Stats:
Temperature:320K(46.85C)
Controller Busy Time:1m0s
Power Cycles:23
Power On Duration:10399h0m0s
Unsafe Shutdowns:19
Error Count:0
Media Errors:0
Read Errors:0
Write Errors:0
Unmap Errors:0
Checksum Errors:0
Error Log Entries:0
Critical Warnings:
Temperature: OK
Available Spare: OK
Device Reliability: OK
Read Only: OK
Volatile Memory Backup: OK
Intel Vendor SMART Attributes:
Program Fail Count:
Normalized:100%
Raw:0
Erase Fail Count:
Normalized:100%
Raw:0
Wear Leveling Count:
Normalized:100%
Min:0
Max:9
Avg:3
End-to-End Error Detection Count:0
CRC Error Count:0
Timed Workload, Media Wear:65535
Timed Workload, Host Read/Write Ratio:65535
Timed Workload, Timer:65535
Thermal Throttle Status:0%
Thermal Throttle Event Count:0
Retry Buffer Overflow Counter:0
PLL Lock Loss Count:0
NAND Bytes Written:222897
Host Bytes Written:71
Selection¶
The next step consists of specifying the devices that should be used by DAOS
in the server configuration YAML file.
The engines
section of the config is a list with details for each DAOS engine
to be started on the host (one engine will be created for each entry in list).
Devices with the same NUMA rating/node/socket should be colocated on a single DAOS engine where possible. See the server configuration file section for more details.
Storage is specified for each engine in a storage
list with each entry in
the list defining an individual storage tier.
Each tier has a class
parameter which defines the storage type.
Typical class values are "dcpm" for PMem (Intel(R) Optane(TM) persistent
memory) and "nvme" for NVMe SSDs.
When persistent memory is unavailable, class may be set to "ram", which
emulates SCM using a ramfs device, and metadata and small objects are
saved to NVMe SSDs using logging and checkpointing.
For class == "dcpm", the following parameters should be populated:
scm_list
should should contain PMem interleaved-set namespaces (e.g./dev/pmem1
). Currently the size of the list is limited to 1.scm_mount
gives the desired local directory to be used as the mount point for DAOS persistent storage mounted on the specified PMem device specified inscm_list
.
For class == "ram", scm_list
is omitted and scm_size
is specified instead.
In this case, the subsequent bdev tiers describe the persistent storage used
for both data and metadata.
scm_size
specifies the amount of RAM dedicated to emulate SCM for holding DAOS metadata. As with SCM, the RAM size must be sufficient to hold metadata to accommodate the size of data tiers. Required metadata to data ratios vary by usage, but typically metadata size will only be a few percent of data. The storage requirements discussion of the ratio between SCM size and the size of NVMe data tiers is relevant, as the required RAM / NVMe ratio will be similar.
For class == "nvme", the following parameters should be populated:
bdev_list
should be populated with NVMe PCI addresses.bdev_roles
optionally specifies a list of roles for this tier. By default, the DAOS server will assign roles to bdev tiers automatically, so the bdev_roles directive is only needed when that assignment doesn't match your use case.
When "dcpm" is used for the first tier, this list should be omitted or specify only "data". Only a single NVMe tier is supported.
When class == "ram" is used, the NVMe tier roles can be one or more of "wal" (write-ahead-log for tracking the changes made by local transactions), "meta" (for persistent metadata and small object storage), or "data" (contents of larger objects). Only the "data" role may be assigned to multiple tiers. If no roles are specified, then the server will assign them. Otherwise all roles must be assigned to a tier.
See the sample configuration file
daos_server.yml
and example configuration files in the
examples
directory for more details.
The default way for DAOS to access NVMe storage is through SPDK via the VFIO user-space driver.
To use an alternative driver with SPDK, set disable_vfio: true
in the global section of the
server config file to fallback to using UIO user-space driver with SPDK instead.
Note
If UIO user-space driver is used instead of VFIO, 'daos_server' needs to be run as root.
If VMD is enabled on a host, its usage will be enabled by default meaning that the bdev_list
device addresses will be interpreted as VMD endpoints and storage scan will report the details of
the physical NVMe backing devices that belong to each VMD endpoint. To disable the use of VMD on a
VMD-enabled host, set disable_vmd: true
in the global section of the config to fallback to using
physical NVMe devices only.
Warning
If upgrading from DAOS 2.0 to a greater version, the old 'enable_vmd' server config file
parameter is no longer honored and instead should be removed (or replaced by
disable_vmd: true
if VMD is to be explicitly disabled).
Otherwise 'daos_server' may fail config validation and not start after an update from 2.0 to a greater version.
Example Configurations¶
To illustrate, assume a cluster with homogeneous hardware configurations that returns the following from scan for each host:
[daos@wolf-72 daos_m]$ dmg -l wolf-7[1-2] storage scan --verbose
-------
wolf-7[1-2]
-------
SCM Namespace Socket ID Capacity
------------- --------- --------
pmem0 0 2.90TB
pmem1 1 2.90TB
NVMe PCI Model FW Revision Socket ID Capacity
-------- ----- ----------- --------- --------
0000:81:00.0 INTEL SSDPED1K750GA E2010325 0 750.00GB
0000:87:00.0 INTEL SSDPEDMD016T4 8DV10171 0 1.56TB
0000:da:00.0 INTEL SSDPED1K750GA E2010325 1 750.00GB
In this situation, the configuration file engines
section could be populated
as follows to establish 2-tier storage:
<snip>
port: 10001
access_points: ["wolf-71"] # <----- updated
<snip>
engines:
-
pinned_numa_node: 0 # run 1st engine on CPU 0
targets: 16 # number of I/O service threads per-engine
nr_xs_helpers: 4 # count of I/O offload threads
fabric_iface: eth0 # network interface to use for this engine
fabric_iface_port: 31416 # network port
log_mask: ERR # debug level to start with the engine with
log_file: /tmp/server1.log # where to store engine logs
storage:
-
class: dcpm # type of first storage tier (SCM)
scm_list: [/dev/pmem0] # <----- updated
scm_mount: /mnt/daos0 # where to mount SCM
-
class: nvme # type of second storage tier (NVMe)
bdev_list: ["0000:87:00.0"] # <----- updated
-
pinned_numa_node: 1
targets: 16
nr_xs_helpers: 4
fabric_iface: eth0
fabric_iface_port: 32416
log_mask: ERR
log_file: /tmp/server2.log
storage:
-
class: dcpm # type of first storage tier (SCM)
scm_list: [/dev/pmem1] # <----- updated
scm_mount: /mnt/daos1 # where to mount SCM
-
class: nvme # type of second storage tier (NVMe)
bdev_list: ["0000:da:00.0"] # <----- updated
<end>
There are a few optional providers that are not built by default. For detailed information, please refer to the DAOS build documentation.
Note
DAOS Control Servers will need to be restarted on all hosts after updates to the server configuration file.
Pick an odd number of hosts in the system and set access_points
to list of that host's
hostname or IP address (don't need to specify port).
This will be the host which bootstraps the DAOS management service (MS).
The support of the optional providers is not guarantee and can be removed without further notification.
Network Configuration¶
Network Scan¶
The dmg
utility supports the network scan
function to display the network
interfaces, related OFI fabric providers and associated NUMA node for each
device.
This information is used to configure the global fabric provider and the unique
local network interface for each I/O engine on the storage nodes.
This section will help you determine what to provide for the provider
,
fabric_iface
and pinned_numa_node
entries in the daos_server.yml
file.
The following commands are typical examples:
$ dmg network scan
$ dmg network scan -p all
$ dmg network scan -p ofi+tcp
$ dmg network scan --provider ofi+verbs
In the early stages when a daos_server
has not yet been fully configured and
lacks a declaration of the system's fabric provider, it may be helpful to view
an unfiltered list of scan results.
Use either of these dmg
commands in the early stages to accomplish
this goal:
$ dmg network scan
$ dmg network scan -p all
Typical network scan results look as follows:
$ dmg network scan
-------
wolf-29
-------
-------------
NUMA Socket 1
-------------
Provider Interfaces
-------- ----------
ofi+tcp ib1
---------
localhost
---------
-------------
NUMA Socket 0
-------------
Provider Interfaces
-------- ----------
ofi+tcp ib0, eth0
-------------
NUMA Socket 1
-------------
Provider Interfaces
-------- ----------
ofi+tcp ib1
Use one of these providers to configure the provider
in the daos_server.yml
.
Only one provider may be specified for the entire DAOS installation.
Client nodes must be capable of communicating to the daos_server
nodes via
the same provider.
Therefore, it is helpful to choose network settings for the daos_server
that
are compatible with the expected client node configuration.
After the daos_server.yml
file has been edited and contains a provider,
subsequent dmg network scan
commands will filter the results based on
that provider.
If it is desired to view an unfiltered list again, issue dmg network
scan -p all
.
Regardless of the provider in the daos_server.yml
file, the results may be
filtered to the specified provider with the command dmg network scan
-p ofi_provider
where ofi_provider
is one of the available providers from
the list.
The results of the network scan may be used to help configure the I/O engines.
Each I/O engine is configured with a unique fabric_iface
and optional
pinned_numa_node
.
The interfaces and NUMA Sockets listed in the scan results map to the
daos_server.yml
fabric_iface
and pinned_numa_node
respectively.
The use of pinned_numa_node
is optional, but recommended for best
performance.
When specified with the value that matches the network interface, the I/O
engine will bind itself to that NUMA node and to cores purely within that NUMA
node.
This configuration yields the fastest access to that network device.
Changing Network Providers¶
Information about the network configuration is stored as metadata on the DAOS storage.
If, after initial deployment, the provider must be changed, it is necessary to
reformat the storage devices using dmg storage format
after the configuration
file has been updated with the new provider.
Provider Testing¶
Then, the fi_pingpong
test can be used to verify that the targeted OFI
provider works fine:
node1$ fi_pingpong -p verbs
node2$ fi_pingpong -p verbs ${IP_ADDRESS_NODE1}
bytes #sent #ack total time MB/sec usec/xfer Mxfers/sec
64 10 =10 1.2k 0.00s 21.69 2.95 0.34
256 10 =10 5k 0.00s 116.36 2.20 0.45
1k 10 =10 20k 0.00s 379.26 2.70 0.37
4k 10 =10 80k 0.00s 1077.89 3.80 0.26
64k 10 =10 1.2m 0.00s 2145.20 30.55 0.03
1m 10 =10 20m 0.00s 8867.45 118.25 0.01
CPU Resources¶
The I/O engine is multi-threaded, and the number of I/O service threads
and helper threads that should be used per engine must be configured
in the engines:
section of the daos_server.yml
file.
The number of I/O service threads is configured with the targets:
setting.
Each storage target manages a fraction of the (interleaved) SCM storage space,
and a fraction of one of the NVMe SSDs that are managed by this engine.
The optimal number of storage targets per engine depends on two conditions:
-
For optimal balance regarding the NVMe space, the number of targets should be an integer multiple of the number of NVMe disks that are configured in the
bdev_list:
of the engine. -
To obtain the maximum SCM performance, a certain number of targets is needed. This is device- and workload-dependent, but around 16 targets usually work well.
It is recommended to also specify a number of I/O offloading threads with the
nr_xs_helpers:
setting. These threads can
improve performance by offloading activities like checksum calculation and
the dispatching of server-side RPCs from the main I/O service threads.
When using EC, it is recommended to configure roughly one offloading thread
per four target threads, for example targets: 16
and nr_xs_helpers: 4
.
The server should have sufficiently many physical cores to support the number of targets plus the additional service threads.
The 'targets:' and 'nr_xs_helpers:' requirement are mandatory, if the number of physical cores are not enough it will fail the starting of the daos engine (notes that 2 cores reserved for system service), or configures with ENV "DAOS_TARGET_OVERSUBSCRIBE=1" to force starting daos engine (possibly hurts performance as multiple XS compete on same core).
Storage Formatting¶
Once the daos_server
has been restarted with the correct storage devices,
network interface, and CPU threads to use, one can move to the format phase.
When daos_server
is started for the first time, it enters "maintenance mode"
and waits for a dmg storage format
call to be issued from the management tool.
This remote call will trigger the formatting of the locally attached storage on
the host for use with DAOS using the parameters defined in the server config file.
dmg -l <host>[,...] storage format
will normally be run on a login
node specifying a hostlist (-l <host>[,...]
) of storage nodes with SCM/PMem
modules and NVMe SSDs installed and prepared.
Upon successful format, DAOS Control Servers will start DAOS I/O engines that have been specified in the server config file.
Successful start-up is indicated by the following on stdout:
DAOS I/O Engine (v2.0.1) process 433456 started on rank 1 with 8 target, 2 helper XS, firstcore 0, host wolf-72.wolf.hpdd.intel.com.
SCM Format¶
When the command is run, the pmem kernel devices created on SCM/PMem regions are formatted and mounted based on the parameters provided in the server config file.
scm_mount
specifies the location of the mountpoint to create.class
can be set toram
to use a tmpfs in the situation that no SCM/PMem is available (scm_size
dictates the size of tmpfs in GB), when set todcpm
the device specified underscm_list
will be mounted atscm_mount
path.
NVMe Format¶
When the command is run, NVMe SSDs are formatted and set up to be used by DAOS based on the parameters provided in the server config file engine storage sections.
class
can be set to nvme
to use actual NVMe devices with SPDK for DAOS
storage.
Other class
values can be used for emulation of NVMe storage as specified
in the server config file.
bdev_list
identifies devices to use with a list of PCI addresses (this can be
populated after viewing results from storage scan
command).
After the format command is run, the path specified by the server configuration
file scm_mount
parameter should be mounted and should contain a file named
daos_nvme.conf
.
The file should describe the devices with PCI addresses as listed in the
bdev_list
parameter of the server config file.
The presence and contents of the file indicate that the specified NVMe SSDs have
been configured correctly for use with DAOS.
The contents of the NVMe SSDs listed in the server configuration file bdev_list
parameter will be reset on format.
Server Format¶
Before the format command is run, no DAOS metadata should exist under the
path specified by scm_mount
parameter in the server configuration file.
After the storage format
command is run, the path specified by the server
configuration file scm_mount
parameter should be mounted and should contain
the necessary DAOS metadata indicating that the server has been formatted.
When starting, daos_server
will skip maintenance mode
and attempt to start
I/O engines if valid DAOS metadata is found in scm_mount
.
Agent Setup¶
This section addresses how to configure the DAOS Agents on the client nodes.
Agent User and Group Setup¶
The daos_agent
daemon runs as an unprivileged user process,
with the username daos_agent
and group name daos_agent
.
If those IDs do not exist at the time the daos-client
RPM is installed,
they will be created during the RPM installation. Refer to
User and Group Management
for details.
Agent Certificate Generation¶
The DAOS security framework relies on SSL certificates to authenticate
the DAOS daemon processes, as well as the DAOS administrators who run the
dmg
command.
Refer to Certificate Generation
for details on creating the necessary certificates.
Note
It is possible to disable the use of certificates for testing purposes.
This should never be done in production environments.
Running in insecure mode will allow arbitrary un-authenticated user processes to access and potentially damage the DAOS storage.
Agent Configuration File¶
The daos_agent
configuration file is parsed when starting the
daos_agent
process. The configuration file location can be specified
on the command line with the -o
option (see daos_agent -h
for usage help).
Otherwise, the default location /etc/daos/daos_agent.yml
will be used.
When running daos_agent
under systemd control, the default location
is used unless the ExecStart
line in the daos_agent.service
file is
modified to include the -o
option.
Parameter descriptions are specified in the sample
daos_agent.yml
file, which will also get installed into /etc/daos/daos_agent.yml
during the installation of the daos-client RPM.
Any option supplied to daos_agent
as a command line option or flag will take
precedence over equivalent configuration file parameter.
The following section lists the format, options, defaults, and descriptions available in the configuration file.
Defining fabric interfaces manually¶
By default, the DAOS Agent automatically detects all fabric interfaces on the client node. It selects an appropriate one for DAOS I/O based on the NUMA node of the client request and the interface type preferences reported by the DAOS management service.
If the DAOS Agent does not detect the fabric interfaces correctly,
the administrator may define them manually in the Agent configuration file.
These fabric_iface
entries must be organized by NUMA node.
If using the verbs provider, the interface domain is also required.
Example:
fabric_ifaces:
-
numa_node: 0
devices:
-
iface: ib0
domain: mlx5_0
-
iface: ib1
domain: mlx5_1
-
numa_node: 1
devices:
-
iface: ib2
domain: mlx5_2
-
iface: ib3
domain: mlx5_3
Agent Startup¶
The DAOS Agent is a standalone application to be run on each client node.
By default, the DAOS Agent will be run as a systemd service.
The DAOS Agent unit file is installed in the correct location
(/usr/lib/systemd/system/daos_agent.service
) during RPM installation.
After the RPM installation, and after the Agent configuration file has been created, the following commands will enable the DAOS Agent to be started at the next reboot, will start it immediately, and will check the status of the Agent after being started:
$ sudo systemctl enable daos_agent.service
$ sudo systemctl start daos_agent.service
$ sudo systemctl status daos_agent.service
If the DAOS Agent fails to start, check the systemd logs for errors:
$ sudo journalctl --unit daos_agent.service
Starting the DAOS Agent with a non-default configuration¶
To start the DAOS Agent from the command line, for example to run with a non-default configuration file, run:
$ daos_agent -o <'path to agent configuration file/daos_agent.yml'> &
If you wish to use systemd with a development build, you must copy the Agent service
file from utils/systemd/
to /usr/lib/systemd/system/
.
Then modify the ExecStart
line to point to your Agent configuration file:
ExecStart=/usr/bin/daos_agent -o <'path to agent configuration file/daos_agent.yml'>
Once the service file is installed and systemctl daemon-reload
has been run to
reload the configuration, the daos_agent
can be started through systemd
as shown above.
Disable Agent Cache (Optional)¶
In certain circumstances (e.g. for DAOS development or system evaluation), it may be desirable to disable the DAOS Agent's caching mechanism in order to avoid stale system information being retained across reformats of a system. The DAOS Agent normally caches a map of rank-to-fabric URI lookups as well as client network configuration data in order to reduce the number of management RPCs required to start an application. When this information becomes stale, the Agent must be restarted in order to repopulate the cache with new information. Alternatively, the caching mechanism may be disabled, with the tradeoff that each application launch will invoke management RPCs in order to obtain system connection information.
To disable the DAOS Agent caching mechanism, set the following environment
variable before starting the daos_agent
process:
DAOS_AGENT_DISABLE_CACHE=true
If running from systemd, add the following to the daos_agent
service file in
the [Service]
section before reloading systemd and restarting the
daos_agent
service:
Environment=DAOS_AGENT_DISABLE_CACHE=true
Multi-user DFuse setup¶
Running a single-user dfuse instance, for example on a compute node, requires no special setup. However configuration is required for allowing multi-user dfuse on a node.
Updating fuse config¶
Multi-user dfuse makes use of the allow_other
fuse mount option which allows requests from users
other than the user running dfuse. For reasons of safety this option is disabled by default for
fuse and must be enabled by root before any user can use it. To allow this then root must add or
uncomment a line in /etc/fuse.conf
to enable the user_allow_other
setting. The daos-client rpm
does not do this automatically. An administrator must set this option on all nodes on which they
want to provide a persistent multi-user dfuse service.