# STDOUT: ---v---v---v---v---v--- ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.9/lib/python3.6/site-packages/ansible executable location = /opt/ansible-2.9/bin/ansible-playbook python version = 3.6.8 (default, Jan 25 2023, 15:03:30) [GCC 8.5.0 20210514 (Red Hat 8.5.0-18)] Using /etc/ansible/ansible.cfg as config file Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_fatals_cache_volume_scsi_generated.yml ************************* 2 plays in /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume_scsi_generated.yml PLAY [Run test tests_fatals_cache_volume.yml for scsi] ************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume_scsi_generated.yml:3 Wednesday 31 May 2023 20:46:50 +0000 (0:00:00.015) 0:00:00.015 ********* ok: [sut] META: ran handlers TASK [Set disk interface for test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume_scsi_generated.yml:8 Wednesday 31 May 2023 20:46:51 +0000 (0:00:00.802) 0:00:00.818 ********* ok: [sut] => { "ansible_facts": { "storage_test_use_interface": "scsi" }, "changed": false } META: ran handlers META: ran handlers PLAY [Test cache volume fatal errors] ****************************************** TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:2 Wednesday 31 May 2023 20:46:51 +0000 (0:00:00.016) 0:00:00.834 ********* ok: [sut] META: ran handlers TASK [Run the role] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:14 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.521) 0:00:01.356 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.023) 0:00:01.379 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.033) 0:00:01.413 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.349) 0:00:01.762 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.045) 0:00:01.808 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.012) 0:00:01.821 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.013) 0:00:01.835 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Wednesday 31 May 2023 20:46:52 +0000 (0:00:00.042) 0:00:01.878 ********* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-blivet-1:3.4.4-1.fc36.noarch", "Installed: python3-blockdev-2.28-2.fc36.x86_64", "Installed: python3-bytesize-2.7-1.fc36.x86_64", "Installed: device-mapper-event-1.02.175-7.fc36.x86_64", "Installed: libblockdev-btrfs-2.28-2.fc36.x86_64", "Installed: lzo-2.10-6.fc36.x86_64", "Installed: device-mapper-event-libs-1.02.175-7.fc36.x86_64", "Installed: sgpio-1.2.0.10-30.fc36.x86_64", "Installed: device-mapper-persistent-data-0.9.0-7.fc36.x86_64", "Installed: python3-pyparted-1:3.12.0-1.fc36.x86_64", "Installed: libblockdev-dm-2.28-2.fc36.x86_64", "Installed: lvm2-2.03.11-7.fc36.x86_64", "Installed: cxl-libs-76.1-1.fc36.x86_64", "Installed: lvm2-libs-2.03.11-7.fc36.x86_64", "Installed: libblockdev-kbd-2.28-2.fc36.x86_64", "Installed: blivet-data-1:3.4.4-1.fc36.noarch", "Installed: libblockdev-lvm-2.28-2.fc36.x86_64", "Installed: libblockdev-mpath-2.28-2.fc36.x86_64", "Installed: libblockdev-nvdimm-2.28-2.fc36.x86_64", "Installed: ndctl-76.1-1.fc36.x86_64", "Installed: lsof-4.94.0-3.fc36.x86_64", "Installed: device-mapper-multipath-0.8.7-9.fc36.x86_64", "Installed: bcache-tools-1.1-2.fc36.x86_64", "Installed: ndctl-libs-76.1-1.fc36.x86_64", "Installed: device-mapper-multipath-libs-0.8.7-9.fc36.x86_64", "Installed: daxctl-libs-76.1-1.fc36.x86_64", "Installed: btrfs-progs-6.2.2-1.fc36.x86_64", "Installed: dmraid-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-events-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-libs-1.0.0.rc16-52.fc36.x86_64", "Installed: libaio-0.3.111-13.fc36.x86_64", "Installed: iniparser-4.1-9.fc36.x86_64" ] } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Wednesday 31 May 2023 20:49:09 +0000 (0:02:16.559) 0:02:18.437 ********* ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Wednesday 31 May 2023 20:49:09 +0000 (0:00:00.028) 0:02:18.465 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Wednesday 31 May 2023 20:49:09 +0000 (0:00:00.026) 0:02:18.492 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Wednesday 31 May 2023 20:49:09 +0000 (0:00:00.758) 0:02:19.250 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Wednesday 31 May 2023 20:49:09 +0000 (0:00:00.047) 0:02:19.298 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Wednesday 31 May 2023 20:49:10 +0000 (0:00:00.022) 0:02:19.320 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Wednesday 31 May 2023 20:49:10 +0000 (0:00:00.013) 0:02:19.334 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Wednesday 31 May 2023 20:49:10 +0000 (0:00:00.022) 0:02:19.357 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Wednesday 31 May 2023 20:51:36 +0000 (0:02:26.357) 0:04:45.714 ********* ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "inactive", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-system-token.service": { "name": "systemd-boot-system-token.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Wednesday 31 May 2023 20:51:38 +0000 (0:00:02.310) 0:04:48.024 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Wednesday 31 May 2023 20:51:38 +0000 (0:00:00.032) 0:04:48.057 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Wednesday 31 May 2023 20:51:38 +0000 (0:00:00.015) 0:04:48.073 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.372) 0:04:48.446 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.019) 0:04:48.465 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.016) 0:04:48.482 ********* ok: [sut] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.019) 0:04:48.501 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.021) 0:04:48.522 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.019) 0:04:48.542 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.017) 0:04:48.559 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.018) 0:04:48.578 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.017) 0:04:48.595 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.016) 0:04:48.612 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.018) 0:04:48.631 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685564517.381066, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.288) 0:04:48.920 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Wednesday 31 May 2023 20:51:39 +0000 (0:00:00.017) 0:04:48.937 ********* ok: [sut] TASK [Mark tasks to be skipped] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:18 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.618) 0:04:49.555 ********* ok: [sut] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Get unused disks] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:25 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.017) 0:04:49.573 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml for sut TASK [Find unused disks in the system] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:2 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.027) 0:04:49.601 ********* ok: [sut] => { "changed": false, "disks": [ "sda", "sdb" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:10 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.301) 0:04:49.902 ********* ok: [sut] => { "ansible_facts": { "unused_disks": [ "sda", "sdb" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:15 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.023) 0:04:49.925 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:20 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.039) 0:04:49.965 ********* ok: [sut] => { "unused_disks": [ "sda", "sdb" ] } TASK [Verify that creating a cached partition volume fails] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:31 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.020) 0:04:49.985 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml for sut TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:4 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.029) 0:04:50.015 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.031) 0:04:50.046 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Wednesday 31 May 2023 20:51:40 +0000 (0:00:00.026) 0:04:50.073 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.355) 0:04:50.428 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.045) 0:04:50.474 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.018) 0:04:50.492 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.019) 0:04:50.511 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.042) 0:04:50.554 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.018) 0:04:50.573 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda", "sdb" ], "name": "sda", "type": "partition", "volumes": [ { "cached": true, "name": "test1", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.024) 0:04:50.597 ********* ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.024) 0:04:50.622 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.018) 0:04:50.640 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.018) 0:04:50.659 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.019) 0:04:50.678 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.019) 0:04:50.698 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.030) 0:04:50.729 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Wednesday 31 May 2023 20:51:41 +0000 (0:00:00.017) 0:04:50.746 ********* fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting and/or devices on disk 'sda' (pool 'sda') in safe mode TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:95 Wednesday 31 May 2023 20:51:44 +0000 (0:00:03.449) 0:04:54.195 ********* fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting and/or devices on disk 'sda' (pool 'sda') in safe mode", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda', 'sdb'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'sda', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': '', 'fs_type': None, 'mount_options': None, 'mount_point': None, 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': None, 'state': 'present', 'type': 'partition', 'cached': True, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Wednesday 31 May 2023 20:51:44 +0000 (0:00:00.023) 0:04:54.219 ********* TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:20 Wednesday 31 May 2023 20:51:44 +0000 (0:00:00.016) 0:04:54.235 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:25 Wednesday 31 May 2023 20:51:44 +0000 (0:00:00.020) 0:04:54.255 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:36 Wednesday 31 May 2023 20:51:44 +0000 (0:00:00.018) 0:04:54.273 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that creating cache on unused disk fails] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:44 Wednesday 31 May 2023 20:51:44 +0000 (0:00:00.018) 0:04:54.292 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml for sut TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:4 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.031) 0:04:54.323 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.028) 0:04:54.351 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.027) 0:04:54.379 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.348) 0:04:54.727 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.044) 0:04:54.772 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.035) 0:04:54.808 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.018) 0:04:54.826 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.040) 0:04:54.867 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.019) 0:04:54.886 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "cache_devices": [ "sdb" ], "cache_size": "4g", "cached": true, "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.024) 0:04:54.911 ********* ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.024) 0:04:54.935 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.018) 0:04:54.954 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.018) 0:04:54.972 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.018) 0:04:54.990 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.019) 0:04:55.010 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.030) 0:04:55.040 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Wednesday 31 May 2023 20:51:45 +0000 (0:00:00.015) 0:04:55.056 ********* fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cache device 'sdb' doesn't seems to be a physical volume or its parent TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:95 Wednesday 31 May 2023 20:51:46 +0000 (0:00:01.217) 0:04:56.273 ********* fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cache device 'sdb' doesn't seems to be a physical volume or its parent", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '5g', 'state': 'present', 'type': 'lvm', 'cached': True, 'cache_devices': ['sdb'], 'cache_mode': None, 'cache_size': '4g', 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Wednesday 31 May 2023 20:51:46 +0000 (0:00:00.024) 0:04:56.297 ********* TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:20 Wednesday 31 May 2023 20:51:47 +0000 (0:00:00.016) 0:04:56.314 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:25 Wednesday 31 May 2023 20:51:47 +0000 (0:00:00.021) 0:04:56.336 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:36 Wednesday 31 May 2023 20:51:47 +0000 (0:00:00.018) 0:04:56.354 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* sut : ok=50 changed=1 unreachable=0 failed=2 skipped=31 rescued=2 ignored=0 Wednesday 31 May 2023 20:51:47 +0000 (0:00:00.009) 0:04:56.364 ********* =============================================================================== linux-system-roles.storage : Make sure required packages are installed - 146.36s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 linux-system-roles.storage : Make sure blivet is available ------------ 136.56s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 3.45s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Get service facts -------------------------- 2.31s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 1.22s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Gathering Facts --------------------------------------------------------- 0.80s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume_scsi_generated.yml:3 linux-system-roles.storage : Get required packages ---------------------- 0.76s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 linux-system-roles.storage : Update facts ------------------------------- 0.62s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Gathering Facts --------------------------------------------------------- 0.52s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_fatals_cache_volume.yml:2 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 0.37s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.36s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.35s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 linux-system-roles.storage : Ensure ansible_facts used by role ---------- 0.35s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Find unused disks in the system ----------------------------------------- 0.30s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:2 ---- linux-system-roles.storage : Retrieve facts for the /etc/crypttab file --- 0.29s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 linux-system-roles.storage : Include the appropriate provider tasks ----- 0.08s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 linux-system-roles.storage : Set platform/version specific variables ---- 0.05s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 linux-system-roles.storage : Define an empty list of pools to be used in testing --- 0.05s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 linux-system-roles.storage : Enable copr repositories if needed --------- 0.05s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 linux-system-roles.storage : Set platform/version specific variables ---- 0.05s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- /opt/ansible-2.9/lib/python3.6/site-packages/ansible/parsing/vault/__init__.py:44: CryptographyDeprecationWarning: Python 3.6 is no longer supported by the Python core team. Therefore, support for it is deprecated in cryptography. The next release of cryptography will remove support for Python 3.6. from cryptography.exceptions import InvalidSignature [DEPRECATION WARNING]: Distribution fedora 36 on host sut should use /usr/bin/python3, but is using /usr/bin/python for backward compatibility with prior Ansible releases. A future Ansible release will default to using the discovered platform python for this host. See https://docs.ansible.com/ansible/ 2.9/reference_appendices/interpreter_discovery.html for more information. This feature will be removed in version 2.12. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. ---^---^---^---^---^---