# STDOUT: ---v---v---v---v---v--- ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.9/lib/python3.6/site-packages/ansible executable location = /opt/ansible-2.9/bin/ansible-playbook python version = 3.6.8 (default, Jan 25 2023, 15:03:30) [GCC 8.5.0 20210514 (Red Hat 8.5.0-18)] Using /etc/ansible/ansible.cfg as config file Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_raid_volume_options.yml **************************************** 1 plays in /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml PLAY [Test raid volume options] ************************************************ TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:2 Thursday 01 June 2023 01:58:01 +0000 (0:00:00.019) 0:00:00.019 ********* ok: [sut] META: ran handlers TASK [Run the role] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:11 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.825) 0:00:00.844 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.035) 0:00:00.880 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.033) 0:00:00.913 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.352) 0:00:01.266 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.047) 0:00:01.314 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.012) 0:00:01.327 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.012) 0:00:01.339 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 01:58:02 +0000 (0:00:00.040) 0:00:01.380 ********* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-blivet-1:3.4.4-1.fc36.noarch", "Installed: python3-blockdev-2.28-2.fc36.x86_64", "Installed: python3-bytesize-2.7-1.fc36.x86_64", "Installed: device-mapper-event-1.02.175-7.fc36.x86_64", "Installed: libblockdev-btrfs-2.28-2.fc36.x86_64", "Installed: lzo-2.10-6.fc36.x86_64", "Installed: device-mapper-event-libs-1.02.175-7.fc36.x86_64", "Installed: sgpio-1.2.0.10-30.fc36.x86_64", "Installed: device-mapper-persistent-data-0.9.0-7.fc36.x86_64", "Installed: python3-pyparted-1:3.12.0-1.fc36.x86_64", "Installed: libblockdev-dm-2.28-2.fc36.x86_64", "Installed: lvm2-2.03.11-7.fc36.x86_64", "Installed: cxl-libs-76.1-1.fc36.x86_64", "Installed: lvm2-libs-2.03.11-7.fc36.x86_64", "Installed: libblockdev-kbd-2.28-2.fc36.x86_64", "Installed: blivet-data-1:3.4.4-1.fc36.noarch", "Installed: libblockdev-lvm-2.28-2.fc36.x86_64", "Installed: libblockdev-mpath-2.28-2.fc36.x86_64", "Installed: libblockdev-nvdimm-2.28-2.fc36.x86_64", "Installed: ndctl-76.1-1.fc36.x86_64", "Installed: lsof-4.94.0-3.fc36.x86_64", "Installed: device-mapper-multipath-0.8.7-9.fc36.x86_64", "Installed: bcache-tools-1.1-2.fc36.x86_64", "Installed: ndctl-libs-76.1-1.fc36.x86_64", "Installed: device-mapper-multipath-libs-0.8.7-9.fc36.x86_64", "Installed: daxctl-libs-76.1-1.fc36.x86_64", "Installed: btrfs-progs-6.2.2-1.fc36.x86_64", "Installed: dmraid-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-events-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-libs-1.0.0.rc16-52.fc36.x86_64", "Installed: libaio-0.3.111-13.fc36.x86_64", "Installed: iniparser-4.1-9.fc36.x86_64" ] } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 01:58:10 +0000 (0:00:07.862) 0:00:09.242 ********* ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 01:58:10 +0000 (0:00:00.024) 0:00:09.267 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 01:58:10 +0000 (0:00:00.023) 0:00:09.291 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 01:58:11 +0000 (0:00:00.566) 0:00:09.857 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 01 June 2023 01:58:11 +0000 (0:00:00.047) 0:00:09.905 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 01 June 2023 01:58:11 +0000 (0:00:00.022) 0:00:09.927 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 01 June 2023 01:58:11 +0000 (0:00:00.013) 0:00:09.941 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 01:58:11 +0000 (0:00:00.022) 0:00:09.963 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 01:58:13 +0000 (0:00:02.418) 0:00:12.382 ********* ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "inactive", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-system-token.service": { "name": "systemd-boot-system-token.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 01:58:16 +0000 (0:00:02.326) 0:00:14.709 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.030) 0:00:14.739 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.015) 0:00:14.755 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.371) 0:00:15.126 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.018) 0:00:15.144 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.015) 0:00:15.160 ********* ok: [sut] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.018) 0:00:15.179 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.018) 0:00:15.197 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.018) 0:00:15.216 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.017) 0:00:15.234 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.031) 0:00:15.265 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.016) 0:00:15.282 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.016) 0:00:15.298 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 01:58:16 +0000 (0:00:00.017) 0:00:15.315 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685584145.694172, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 01:58:17 +0000 (0:00:00.263) 0:00:15.579 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 01:58:17 +0000 (0:00:00.016) 0:00:15.596 ********* ok: [sut] TASK [Mark tasks to be skipped] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:15 Thursday 01 June 2023 01:58:17 +0000 (0:00:00.621) 0:00:16.217 ********* ok: [sut] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Get unused disks] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:22 Thursday 01 June 2023 01:58:17 +0000 (0:00:00.016) 0:00:16.234 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml for sut TASK [Find unused disks in the system] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:2 Thursday 01 June 2023 01:58:17 +0000 (0:00:00.025) 0:00:16.259 ********* ok: [sut] => { "changed": false, "disks": [ "sda", "sdb", "sdc" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:10 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.270) 0:00:16.530 ********* ok: [sut] => { "ansible_facts": { "unused_disks": [ "sda", "sdb", "sdc" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:15 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.021) 0:00:16.552 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:20 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.019) 0:00:16.571 ********* ok: [sut] => { "unused_disks": [ "sda", "sdb", "sdc" ] } TASK [Create a RAID0 device mounted on "/opt/test1"] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:28 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.021) 0:00:16.593 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.033) 0:00:16.626 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.024) 0:00:16.651 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.345) 0:00:16.997 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.044) 0:00:17.041 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.017) 0:00:17.058 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.016) 0:00:17.075 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.035) 0:00:17.111 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.019) 0:00:17.130 ********* ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.019) 0:00:17.149 ********* ok: [sut] => { "storage_volumes": [ { "disks": [ "sda", "sdb", "sdc" ], "mount_point": "/opt/test1", "name": "test1", "raid_device_count": 2, "raid_level": "raid1", "raid_metadata_version": "1.0", "raid_spare_count": 1, "state": "present", "type": "raid" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.021) 0:00:17.171 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.018) 0:00:17.190 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.017) 0:00:17.207 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.017) 0:00:17.224 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.017) 0:00:17.242 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.032) 0:00:17.274 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 01:58:18 +0000 (0:00:00.015) 0:00:17.289 ********* changed: [sut] => { "actions": [ { "action": "create format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdc1", "fs_type": null }, { "action": "create format", "device": "/dev/sdc1", "fs_type": "mdmember" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "mdmember" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "mdmember" }, { "action": "create device", "device": "/dev/md/test1", "fs_type": null }, { "action": "create format", "device": "/dev/md/test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/md/test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "state": "mounted" } ], "packages": [ "mdadm", "e2fsprogs", "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/md/test1", "_kernel_device": "/dev/md127", "_mount_id": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "_raw_device": "/dev/md/test1", "_raw_kernel_device": "/dev/md127", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": 2, "raid_level": "raid1", "raid_metadata_version": "1.0", "raid_spare_count": 1, "raid_stripe_size": null, "size": 0, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "raid", "vdo_pool_size": null } ] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 01:58:28 +0000 (0:00:10.187) 0:00:27.476 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 01:58:29 +0000 (0:00:00.018) 0:00:27.495 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 01:58:29 +0000 (0:00:00.015) 0:00:27.511 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdc1", "fs_type": null }, { "action": "create format", "device": "/dev/sdc1", "fs_type": "mdmember" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "mdmember" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "mdmember" }, { "action": "create device", "device": "/dev/md/test1", "fs_type": null }, { "action": "create format", "device": "/dev/md/test1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/md/test1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "state": "mounted" } ], "packages": [ "mdadm", "e2fsprogs", "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/md/test1", "_kernel_device": "/dev/md127", "_mount_id": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "_raw_device": "/dev/md/test1", "_raw_kernel_device": "/dev/md127", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": 2, "raid_level": "raid1", "raid_metadata_version": "1.0", "raid_spare_count": 1, "raid_stripe_size": null, "size": 0, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "raid", "vdo_pool_size": null } ] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 01:58:29 +0000 (0:00:00.022) 0:00:27.533 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 01:58:29 +0000 (0:00:00.019) 0:00:27.552 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/md/test1", "_kernel_device": "/dev/md127", "_mount_id": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "_raw_device": "/dev/md/test1", "_raw_kernel_device": "/dev/md127", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": 2, "raid_level": "raid1", "raid_metadata_version": "1.0", "raid_spare_count": 1, "raid_stripe_size": null, "size": 0, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "raid", "vdo_pool_size": null } ] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 01:58:29 +0000 (0:00:00.020) 0:00:27.573 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 01:58:29 +0000 (0:00:00.017) 0:00:27.591 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 01:58:30 +0000 (0:00:00.907) 0:00:28.498 ********* changed: [sut] => (item={'src': 'UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 01:58:31 +0000 (0:00:01.047) 0:00:29.545 ********* skipping: [sut] => (item={'src': 'UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 01:58:31 +0000 (0:00:00.025) 0:00:29.571 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 01:58:32 +0000 (0:00:01.069) 0:00:30.641 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685584145.694172, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 01:58:32 +0000 (0:00:00.260) 0:00:30.901 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 01:58:32 +0000 (0:00:00.016) 0:00:30.917 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:43 Thursday 01 June 2023 01:58:33 +0000 (0:00:00.825) 0:00:31.743 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 01:58:33 +0000 (0:00:00.031) 0:00:31.775 ********* skipping: [sut] => {} TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 01:58:33 +0000 (0:00:00.020) 0:00:31.795 ********* ok: [sut] => { "_storage_volumes_list": [ { "_device": "/dev/md/test1", "_kernel_device": "/dev/md127", "_mount_id": "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071", "_raw_device": "/dev/md/test1", "_raw_kernel_device": "/dev/md127", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": 2, "raid_level": "raid1", "raid_metadata_version": "1.0", "raid_spare_count": 1, "raid_stripe_size": null, "size": 0, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "raid", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 01:58:33 +0000 (0:00:00.024) 0:00:31.819 ********* ok: [sut] => { "changed": false, "info": { "/dev/md/test1": { "fstype": "xfs", "label": "", "name": "/dev/md/test1", "size": "10G", "type": "raid1", "uuid": "73518b25-b92c-40b6-ad0b-c8fb2780f071" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "linux_raid_member", "label": "test1", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "8bcad0bf-21d7-686c-22b4-10b791f3fcef" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "linux_raid_member", "label": "test1", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "8bcad0bf-21d7-686c-22b4-10b791f3fcef" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "linux_raid_member", "label": "test1", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "8bcad0bf-21d7-686c-22b4-10b791f3fcef" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 01:58:33 +0000 (0:00:00.285) 0:00:32.104 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003770", "end": "2023-06-01 01:58:33.870095", "rc": 0, "start": "2023-06-01 01:58:33.866325" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 01:58:33 +0000 (0:00:00.300) 0:00:32.405 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003878", "end": "2023-06-01 01:58:34.211362", "failed_when_result": false, "rc": 0, "start": "2023-06-01 01:58:34.207484" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.333) 0:00:32.738 ********* TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.016) 0:00:32.755 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.038) 0:00:32.794 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.022) 0:00:32.816 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.106) 0:00:32.923 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/md127" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.023) 0:00:32.946 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2590293, "block_size": 4096, "block_total": 2616800, "block_used": 26507, "device": "/dev/md127", "fstype": "xfs", "inode_available": 5238717, "inode_total": 5238720, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10609840128, "size_total": 10718412800, "uuid": "73518b25-b92c-40b6-ad0b-c8fb2780f071" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2590293, "block_size": 4096, "block_total": 2616800, "block_used": 26507, "device": "/dev/md127", "fstype": "xfs", "inode_available": 5238717, "inode_total": 5238720, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10609840128, "size_total": 10718412800, "uuid": "73518b25-b92c-40b6-ad0b-c8fb2780f071" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.026) 0:00:32.972 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.019) 0:00:32.992 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.023) 0:00:33.015 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.020) 0:00:33.036 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.021) 0:00:33.057 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.018) 0:00:33.076 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.019) 0:00:33.095 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.022) 0:00:33.118 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.017) 0:00:33.136 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.018) 0:00:33.154 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.019) 0:00:33.174 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.016) 0:00:33.191 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "UUID=73518b25-b92c-40b6-ad0b-c8fb2780f071 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.035) 0:00:33.226 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.022) 0:00:33.249 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.021) 0:00:33.271 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.019) 0:00:33.290 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.017) 0:00:33.307 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.021) 0:00:33.329 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 01:58:34 +0000 (0:00:00.022) 0:00:33.351 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685584710.9826314, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685584708.87363, "dev": 5, "device_type": 2431, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 852, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685584708.87363, "nlink": 1, "path": "/dev/md/test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.215) 0:00:33.567 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.023) 0:00:33.590 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.018) 0:00:33.608 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.023) 0:00:33.632 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "raid" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.020) 0:00:33.652 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "raid1" }, "changed": false } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.023) 0:00:33.676 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.020) 0:00:33.696 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 01:58:35 +0000 (0:00:00.018) 0:00:33.715 ********* fatal: [sut]: FAILED! => { "changed": false, "results": [] } MSG: Failed to download packages: Curl error (23): Failed writing received data to disk/application for https://mirrors.fedoraproject.org/metalink?repo=fedora-36&arch=x86_64 [Failure writing output to destination] to retry, use: --limit @/tmp/tests_raid_volume_options.retry PLAY RECAP ********************************************************************* sut : ok=77 changed=3 unreachable=0 failed=1 skipped=36 rescued=0 ignored=0 Thursday 01 June 2023 01:58:38 +0000 (0:00:03.055) 0:00:36.770 ********* =============================================================================== linux-system-roles.storage : Manage the pools and volumes to match the specified state -- 10.19s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Make sure blivet is available -------------- 7.86s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Ensure cryptsetup is present -------------------------------------------- 3.06s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 linux-system-roles.storage : Make sure required packages are installed --- 2.42s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 linux-system-roles.storage : Get service facts -------------------------- 2.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab --- 1.07s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 linux-system-roles.storage : Set up new/current mounts ------------------ 1.05s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab --- 0.91s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 linux-system-roles.storage : Update facts ------------------------------- 0.83s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Gathering Facts --------------------------------------------------------- 0.83s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_raid_volume_options.yml:2 linux-system-roles.storage : Update facts ------------------------------- 0.62s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 linux-system-roles.storage : Get required packages ---------------------- 0.57s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 0.37s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.35s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.35s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Read the /etc/crypttab file --------------------------------------------- 0.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Read the /etc/fstab file for volume existence --------------------------- 0.30s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Collect info about the volumes. ----------------------------------------- 0.29s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Find unused disks in the system ----------------------------------------- 0.27s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:2 ---- linux-system-roles.storage : Retrieve facts for the /etc/crypttab file --- 0.26s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- /opt/ansible-2.9/lib/python3.6/site-packages/ansible/parsing/vault/__init__.py:44: CryptographyDeprecationWarning: Python 3.6 is no longer supported by the Python core team. Therefore, support for it is deprecated in cryptography. The next release of cryptography will remove support for Python 3.6. from cryptography.exceptions import InvalidSignature [DEPRECATION WARNING]: Distribution fedora 36 on host sut should use /usr/bin/python3, but is using /usr/bin/python for backward compatibility with prior Ansible releases. A future Ansible release will default to using the discovered platform python for this host. See https://docs.ansible.com/ansible/ 2.9/reference_appendices/interpreter_discovery.html for more information. This feature will be removed in version 2.12. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. ---^---^---^---^---^---