# STDOUT: ---v---v---v---v---v--- ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.9/lib/python3.6/site-packages/ansible executable location = /opt/ansible-2.9/bin/ansible-playbook python version = 3.6.8 (default, Jan 25 2023, 15:03:30) [GCC 8.5.0 20210514 (Red Hat 8.5.0-18)] Using /etc/ansible/ansible.cfg as config file Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_null_raid_pool.yml ********************************************* 1 plays in /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml PLAY [Test null raid pool] ***************************************************** TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:2 Thursday 01 June 2023 01:27:07 +0000 (0:00:00.016) 0:00:00.016 ********* ok: [sut] META: ran handlers TASK [Run the role] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:14 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.798) 0:00:00.814 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.034) 0:00:00.849 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.032) 0:00:00.882 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.341) 0:00:01.223 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.045) 0:00:01.269 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.013) 0:00:01.282 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.012) 0:00:01.294 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 01:27:08 +0000 (0:00:00.042) 0:00:01.336 ********* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-blivet-1:3.4.4-1.fc36.noarch", "Installed: python3-blockdev-2.28-2.fc36.x86_64", "Installed: python3-bytesize-2.7-1.fc36.x86_64", "Installed: device-mapper-event-1.02.175-7.fc36.x86_64", "Installed: libblockdev-btrfs-2.28-2.fc36.x86_64", "Installed: lzo-2.10-6.fc36.x86_64", "Installed: device-mapper-event-libs-1.02.175-7.fc36.x86_64", "Installed: sgpio-1.2.0.10-30.fc36.x86_64", "Installed: device-mapper-persistent-data-0.9.0-7.fc36.x86_64", "Installed: python3-pyparted-1:3.12.0-1.fc36.x86_64", "Installed: libblockdev-dm-2.28-2.fc36.x86_64", "Installed: lvm2-2.03.11-7.fc36.x86_64", "Installed: cxl-libs-76.1-1.fc36.x86_64", "Installed: lvm2-libs-2.03.11-7.fc36.x86_64", "Installed: libblockdev-kbd-2.28-2.fc36.x86_64", "Installed: blivet-data-1:3.4.4-1.fc36.noarch", "Installed: libblockdev-lvm-2.28-2.fc36.x86_64", "Installed: libblockdev-mpath-2.28-2.fc36.x86_64", "Installed: libblockdev-nvdimm-2.28-2.fc36.x86_64", "Installed: ndctl-76.1-1.fc36.x86_64", "Installed: lsof-4.94.0-3.fc36.x86_64", "Installed: device-mapper-multipath-0.8.7-9.fc36.x86_64", "Installed: bcache-tools-1.1-2.fc36.x86_64", "Installed: ndctl-libs-76.1-1.fc36.x86_64", "Installed: device-mapper-multipath-libs-0.8.7-9.fc36.x86_64", "Installed: daxctl-libs-76.1-1.fc36.x86_64", "Installed: btrfs-progs-6.2.2-1.fc36.x86_64", "Installed: dmraid-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-events-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-libs-1.0.0.rc16-52.fc36.x86_64", "Installed: libaio-0.3.111-13.fc36.x86_64", "Installed: iniparser-4.1-9.fc36.x86_64" ] } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 01:27:17 +0000 (0:00:08.942) 0:00:10.279 ********* ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 01:27:17 +0000 (0:00:00.025) 0:00:10.304 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 01:27:17 +0000 (0:00:00.025) 0:00:10.330 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 01:27:18 +0000 (0:00:00.636) 0:00:10.967 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 01 June 2023 01:27:18 +0000 (0:00:00.046) 0:00:11.014 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 01 June 2023 01:27:18 +0000 (0:00:00.023) 0:00:11.037 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 01 June 2023 01:27:18 +0000 (0:00:00.014) 0:00:11.052 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 01:27:18 +0000 (0:00:00.022) 0:00:11.074 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 01:27:21 +0000 (0:00:02.338) 0:00:13.412 ********* ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "inactive", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-system-token.service": { "name": "systemd-boot-system-token.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 01:27:23 +0000 (0:00:02.191) 0:00:15.604 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.030) 0:00:15.634 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.015) 0:00:15.650 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.356) 0:00:16.006 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.017) 0:00:16.024 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.014) 0:00:16.039 ********* ok: [sut] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.018) 0:00:16.057 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.018) 0:00:16.076 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.018) 0:00:16.094 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.017) 0:00:16.111 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.017) 0:00:16.129 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.015) 0:00:16.144 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.015) 0:00:16.160 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 01:27:23 +0000 (0:00:00.016) 0:00:16.177 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685582302.9289844, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 01:27:24 +0000 (0:00:00.271) 0:00:16.449 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 01:27:24 +0000 (0:00:00.016) 0:00:16.465 ********* ok: [sut] TASK [Mark tasks to be skipped] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:18 Thursday 01 June 2023 01:27:24 +0000 (0:00:00.597) 0:00:17.063 ********* ok: [sut] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Get unused disks] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:25 Thursday 01 June 2023 01:27:24 +0000 (0:00:00.016) 0:00:17.079 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml for sut TASK [Find unused disks in the system] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:2 Thursday 01 June 2023 01:27:24 +0000 (0:00:00.025) 0:00:17.104 ********* ok: [sut] => { "changed": false, "disks": [ "sda", "sdb" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:10 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.259) 0:00:17.364 ********* ok: [sut] => { "ansible_facts": { "unused_disks": [ "sda", "sdb" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:15 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.021) 0:00:17.385 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:20 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.017) 0:00:17.403 ********* ok: [sut] => { "unused_disks": [ "sda", "sdb" ] } TASK [Get existing raids (before run)] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:31 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.018) 0:00:17.422 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/proc/mdstat" ], "delta": "0:00:00.003436", "end": "2023-06-01 01:27:25.281669", "rc": 0, "start": "2023-06-01 01:27:25.278233" } STDOUT: Personalities : unused devices: TASK [Check that raid_level null does not create raid] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:36 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.247) 0:00:17.670 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.034) 0:00:17.704 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.025) 0:00:17.730 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.334) 0:00:18.064 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.043) 0:00:18.107 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.016) 0:00:18.124 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.016) 0:00:18.140 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.036) 0:00:18.177 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.017) 0:00:18.194 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda", "sdb" ], "name": "vg1", "raid_level": "null", "state": "present", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "2g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.021) 0:00:18.215 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.019) 0:00:18.234 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.018) 0:00:18.253 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.018) 0:00:18.272 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.017) 0:00:18.289 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.017) 0:00:18.307 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.030) 0:00:18.338 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 01:27:25 +0000 (0:00:00.015) 0:00:18.353 ********* changed: [sut] => { "actions": [ { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/vg1-lv1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": "null", "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 01:27:30 +0000 (0:00:04.692) 0:00:23.045 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 01:27:30 +0000 (0:00:00.019) 0:00:23.064 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 01:27:30 +0000 (0:00:00.015) 0:00:23.079 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/vg1-lv1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": "null", "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 01:27:30 +0000 (0:00:00.022) 0:00:23.102 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": "null", "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 01:27:30 +0000 (0:00:00.020) 0:00:23.123 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 01:27:30 +0000 (0:00:00.018) 0:00:23.141 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 01:27:30 +0000 (0:00:00.017) 0:00:23.158 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 01:27:31 +0000 (0:00:00.775) 0:00:23.933 ********* changed: [sut] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 01:27:32 +0000 (0:00:00.683) 0:00:24.616 ********* skipping: [sut] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 01:27:32 +0000 (0:00:00.025) 0:00:24.642 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 01:27:32 +0000 (0:00:00.610) 0:00:25.253 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685582302.9289844, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 01:27:33 +0000 (0:00:00.211) 0:00:25.465 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 01:27:33 +0000 (0:00:00.016) 0:00:25.482 ********* ok: [sut] TASK [Get existing raids (after run)] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:51 Thursday 01 June 2023 01:27:33 +0000 (0:00:00.607) 0:00:26.089 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/proc/mdstat" ], "delta": "0:00:00.003379", "end": "2023-06-01 01:27:33.901428", "rc": 0, "start": "2023-06-01 01:27:33.898049" } STDOUT: Personalities : unused devices: TASK [Cleanup] ***************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:57 Thursday 01 June 2023 01:27:33 +0000 (0:00:00.199) 0:00:26.289 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 01:27:33 +0000 (0:00:00.036) 0:00:26.325 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 01:27:33 +0000 (0:00:00.026) 0:00:26.351 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.332) 0:00:26.684 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.045) 0:00:26.729 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.018) 0:00:26.747 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.017) 0:00:26.765 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.037) 0:00:26.803 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.019) 0:00:26.822 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda", "sdb" ], "name": "vg1", "raid_level": "null", "state": "absent", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "2g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.021) 0:00:26.844 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.019) 0:00:26.864 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.018) 0:00:26.882 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.018) 0:00:26.900 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.018) 0:00:26.919 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.018) 0:00:26.937 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.030) 0:00:26.968 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 01:27:34 +0000 (0:00:00.016) 0:00:26.985 ********* changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "destroy device", "device": "/dev/vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdb1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": "null", "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 01:27:37 +0000 (0:00:03.220) 0:00:30.205 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 01:27:37 +0000 (0:00:00.018) 0:00:30.224 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 01:27:37 +0000 (0:00:00.015) 0:00:30.239 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "destroy device", "device": "/dev/vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdb1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": "null", "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 01:27:37 +0000 (0:00:00.022) 0:00:30.262 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": "null", "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "2g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 01:27:37 +0000 (0:00:00.020) 0:00:30.282 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 01:27:37 +0000 (0:00:00.019) 0:00:30.302 ********* changed: [sut] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv1" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 01:27:38 +0000 (0:00:00.216) 0:00:30.519 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 01:27:38 +0000 (0:00:00.603) 0:00:31.122 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 01:27:38 +0000 (0:00:00.020) 0:00:31.142 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 01:27:38 +0000 (0:00:00.019) 0:00:31.162 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 01:27:39 +0000 (0:00:00.597) 0:00:31.759 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685582302.9289844, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 01:27:39 +0000 (0:00:00.208) 0:00:31.968 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 01:27:39 +0000 (0:00:00.018) 0:00:31.986 ********* ok: [sut] TASK [Compare mdstat results] ************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:72 Thursday 01 June 2023 01:27:40 +0000 (0:00:00.607) 0:00:32.593 ********* ok: [sut] => { "changed": false } MSG: All assertions passed META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* sut : ok=65 changed=5 unreachable=0 failed=0 skipped=35 rescued=0 ignored=0 Thursday 01 June 2023 01:27:40 +0000 (0:00:00.012) 0:00:32.606 ********* =============================================================================== linux-system-roles.storage : Make sure blivet is available -------------- 8.94s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 4.69s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 3.22s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Make sure required packages are installed --- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 linux-system-roles.storage : Get service facts -------------------------- 2.19s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Gathering Facts --------------------------------------------------------- 0.80s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_null_raid_pool.yml:2 linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab --- 0.78s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 linux-system-roles.storage : Set up new/current mounts ------------------ 0.68s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 linux-system-roles.storage : Get required packages ---------------------- 0.64s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab --- 0.61s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 linux-system-roles.storage : Update facts ------------------------------- 0.61s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 linux-system-roles.storage : Update facts ------------------------------- 0.61s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab --- 0.60s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 linux-system-roles.storage : Update facts ------------------------------- 0.60s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab --- 0.60s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 0.36s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 linux-system-roles.storage : Retrieve facts for the /etc/crypttab file --- 0.27s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- /opt/ansible-2.9/lib/python3.6/site-packages/ansible/parsing/vault/__init__.py:44: CryptographyDeprecationWarning: Python 3.6 is no longer supported by the Python core team. Therefore, support for it is deprecated in cryptography. The next release of cryptography will remove support for Python 3.6. from cryptography.exceptions import InvalidSignature [DEPRECATION WARNING]: Distribution fedora 36 on host sut should use /usr/bin/python3, but is using /usr/bin/python for backward compatibility with prior Ansible releases. A future Ansible release will default to using the discovered platform python for this host. See https://docs.ansible.com/ansible/ 2.9/reference_appendices/interpreter_discovery.html for more information. This feature will be removed in version 2.12. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. ---^---^---^---^---^---