# STDOUT: ---v---v---v---v---v--- ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.9/lib/python3.6/site-packages/ansible executable location = /opt/ansible-2.9/bin/ansible-playbook python version = 3.6.8 (default, Jan 25 2023, 15:03:30) [GCC 8.5.0 20210514 (Red Hat 8.5.0-18)] Using /etc/ansible/ansible.cfg as config file Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_resize.yml ***************************************************** 1 plays in /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml PLAY [Test resize] ************************************************************* TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:2 Thursday 01 June 2023 02:43:25 +0000 (0:00:00.017) 0:00:00.017 ********* ok: [sut] META: ran handlers TASK [Run the role] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:21 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.789) 0:00:00.807 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.024) 0:00:00.832 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.033) 0:00:00.865 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.341) 0:00:01.206 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.046) 0:00:01.253 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.012) 0:00:01.265 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.012) 0:00:01.278 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:43:26 +0000 (0:00:00.040) 0:00:01.319 ********* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: python3-blivet-1:3.4.4-1.fc36.noarch", "Installed: python3-blockdev-2.28-2.fc36.x86_64", "Installed: python3-bytesize-2.7-1.fc36.x86_64", "Installed: device-mapper-event-1.02.175-7.fc36.x86_64", "Installed: libblockdev-btrfs-2.28-2.fc36.x86_64", "Installed: lzo-2.10-6.fc36.x86_64", "Installed: device-mapper-event-libs-1.02.175-7.fc36.x86_64", "Installed: sgpio-1.2.0.10-30.fc36.x86_64", "Installed: device-mapper-persistent-data-0.9.0-7.fc36.x86_64", "Installed: python3-pyparted-1:3.12.0-1.fc36.x86_64", "Installed: libblockdev-dm-2.28-2.fc36.x86_64", "Installed: lvm2-2.03.11-7.fc36.x86_64", "Installed: cxl-libs-76.1-1.fc36.x86_64", "Installed: lvm2-libs-2.03.11-7.fc36.x86_64", "Installed: libblockdev-kbd-2.28-2.fc36.x86_64", "Installed: blivet-data-1:3.4.4-1.fc36.noarch", "Installed: libblockdev-lvm-2.28-2.fc36.x86_64", "Installed: libblockdev-mpath-2.28-2.fc36.x86_64", "Installed: libblockdev-nvdimm-2.28-2.fc36.x86_64", "Installed: ndctl-76.1-1.fc36.x86_64", "Installed: lsof-4.94.0-3.fc36.x86_64", "Installed: device-mapper-multipath-0.8.7-9.fc36.x86_64", "Installed: bcache-tools-1.1-2.fc36.x86_64", "Installed: ndctl-libs-76.1-1.fc36.x86_64", "Installed: device-mapper-multipath-libs-0.8.7-9.fc36.x86_64", "Installed: daxctl-libs-76.1-1.fc36.x86_64", "Installed: btrfs-progs-6.2.2-1.fc36.x86_64", "Installed: dmraid-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-events-1.0.0.rc16-52.fc36.x86_64", "Installed: dmraid-libs-1.0.0.rc16-52.fc36.x86_64", "Installed: libaio-0.3.111-13.fc36.x86_64", "Installed: iniparser-4.1-9.fc36.x86_64" ] } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:43:34 +0000 (0:00:07.581) 0:00:08.900 ********* ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:43:34 +0000 (0:00:00.025) 0:00:08.926 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:43:34 +0000 (0:00:00.023) 0:00:08.949 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:43:34 +0000 (0:00:00.596) 0:00:09.546 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 01 June 2023 02:43:35 +0000 (0:00:00.045) 0:00:09.592 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 01 June 2023 02:43:35 +0000 (0:00:00.021) 0:00:09.613 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 01 June 2023 02:43:35 +0000 (0:00:00.013) 0:00:09.627 ********* skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:43:35 +0000 (0:00:00.021) 0:00:09.649 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:43:37 +0000 (0:00:02.340) 0:00:11.989 ********* ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dmraid-activation.service": { "name": "dmraid-activation.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "lvm2-pvscan@.service": { "name": "lvm2-pvscan@.service", "source": "systemd", "state": "unknown", "status": "static" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-convert.service": { "name": "nfs-convert.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "inactive", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-system-token.service": { "name": "systemd-boot-system-token.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-networkd-wait-online.service": { "name": "systemd-networkd-wait-online.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-networkd.service": { "name": "systemd-networkd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "unbound-anchor.service": { "name": "unbound-anchor.service", "source": "systemd", "state": "stopped", "status": "static" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:43:39 +0000 (0:00:02.191) 0:00:14.180 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:43:39 +0000 (0:00:00.032) 0:00:14.212 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:43:39 +0000 (0:00:00.015) 0:00:14.227 ********* ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.365) 0:00:14.593 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.017) 0:00:14.611 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.017) 0:00:14.629 ********* ok: [sut] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.020) 0:00:14.649 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.019) 0:00:14.669 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.018) 0:00:14.687 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.016) 0:00:14.704 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.017) 0:00:14.722 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.016) 0:00:14.739 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.017) 0:00:14.756 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.036) 0:00:14.792 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.265) 0:00:15.058 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:43:40 +0000 (0:00:00.016) 0:00:15.075 ********* ok: [sut] TASK [Mark tasks to be skipped] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:25 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.610) 0:00:15.686 ********* ok: [sut] => { "ansible_facts": { "storage_skip_checks": [ "blivet_available", "packages_installed", "service_facts" ] }, "changed": false } TASK [Get unused disks] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:32 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.017) 0:00:15.703 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml for sut TASK [Find unused disks in the system] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:2 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.028) 0:00:15.731 ********* ok: [sut] => { "changed": false, "disks": [ "sda" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:10 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.267) 0:00:15.999 ********* ok: [sut] => { "ansible_facts": { "unused_disks": [ "sda" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:15 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.023) 0:00:16.022 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/get_unused_disk.yml:20 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.024) 0:00:16.046 ********* ok: [sut] => { "unused_disks": [ "sda" ] } TASK [Create one LVM logical volume under one volume group with size 5g] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:40 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.025) 0:00:16.071 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.044) 0:00:16.116 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.037) 0:00:16.153 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.335) 0:00:16.489 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.047) 0:00:16.536 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:43:41 +0000 (0:00:00.017) 0:00:16.554 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.017) 0:00:16.571 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.039) 0:00:16.610 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.017) 0:00:16.628 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.021) 0:00:16.649 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.018) 0:00:16.667 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.018) 0:00:16.686 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.018) 0:00:16.704 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.017) 0:00:16.722 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.017) 0:00:16.740 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.029) 0:00:16.770 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:43:42 +0000 (0:00:00.015) 0:00:16.785 ********* changed: [sut] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:43:45 +0000 (0:00:03.634) 0:00:20.419 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:43:45 +0000 (0:00:00.046) 0:00:20.466 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:43:45 +0000 (0:00:00.018) 0:00:20.485 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:43:45 +0000 (0:00:00.023) 0:00:20.508 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:43:45 +0000 (0:00:00.022) 0:00:20.530 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:43:45 +0000 (0:00:00.021) 0:00:20.552 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:43:46 +0000 (0:00:00.019) 0:00:20.571 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:43:46 +0000 (0:00:00.743) 0:00:21.315 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:43:47 +0000 (0:00:00.290) 0:00:21.605 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:43:47 +0000 (0:00:00.024) 0:00:21.630 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:43:47 +0000 (0:00:00.597) 0:00:22.227 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:43:47 +0000 (0:00:00.206) 0:00:22.434 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:43:47 +0000 (0:00:00.017) 0:00:22.451 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:57 Thursday 01 June 2023 02:43:48 +0000 (0:00:00.616) 0:00:23.068 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:43:48 +0000 (0:00:00.031) 0:00:23.100 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:43:48 +0000 (0:00:00.024) 0:00:23.124 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:43:48 +0000 (0:00:00.018) 0:00:23.143 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "ldgszG-feN2-QvqV-idFl-en1N-ZwC8-lFZ2Mr" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:43:48 +0000 (0:00:00.291) 0:00:23.434 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003380", "end": "2023-06-01 02:43:49.116492", "rc": 0, "start": "2023-06-01 02:43:49.113112" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.269) 0:00:23.703 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.004932", "end": "2023-06-01 02:43:49.320455", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:43:49.315523" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.210) 0:00:23.914 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.040) 0:00:23.955 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.018) 0:00:23.973 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.041) 0:00:24.015 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.024) 0:00:24.040 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.275) 0:00:24.316 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.022) 0:00:24.338 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.024) 0:00:24.363 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.024) 0:00:24.387 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.021) 0:00:24.409 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.021) 0:00:24.431 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.017) 0:00:24.449 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.025) 0:00:24.474 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.055) 0:00:24.530 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:43:49 +0000 (0:00:00.019) 0:00:24.550 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.567 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.585 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.603 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.620 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.019) 0:00:24.639 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.657 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.675 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.039) 0:00:24.714 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.040) 0:00:24.755 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:24.773 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:24.791 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:24.809 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:24.828 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.019) 0:00:24.847 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:24.865 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:24.884 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.038) 0:00:24.922 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.038) 0:00:24.961 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.979 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:24.997 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.014 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.032 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.039) 0:00:25.072 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.021) 0:00:25.093 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.021) 0:00:25.115 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.037) 0:00:25.152 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.021) 0:00:25.174 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.021) 0:00:25.196 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:25.214 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:25.232 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.019) 0:00:25.252 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.270 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.287 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.041) 0:00:25.328 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.040) 0:00:25.369 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:25.387 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.404 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:25.423 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.441 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.019) 0:00:25.460 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.018) 0:00:25.478 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.016) 0:00:25.495 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.017) 0:00:25.512 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:43:50 +0000 (0:00:00.036) 0:00:25.548 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.023) 0:00:25.572 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.113) 0:00:25.685 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.023) 0:00:25.708 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1199010, "block_size": 4096, "block_total": 1268648, "block_used": 69638, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4911144960, "size_total": 5196382208, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1199010, "block_size": 4096, "block_total": 1268648, "block_used": 69638, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4911144960, "size_total": 5196382208, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.030) 0:00:25.739 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.019) 0:00:25.759 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.022) 0:00:25.782 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.021) 0:00:25.804 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.018) 0:00:25.822 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.020) 0:00:25.842 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.018) 0:00:25.861 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.022) 0:00:25.883 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.018) 0:00:25.902 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.018) 0:00:25.921 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.019) 0:00:25.940 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.020) 0:00:25.961 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.038) 0:00:26.000 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.023) 0:00:26.024 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.022) 0:00:26.046 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.018) 0:00:26.065 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.020) 0:00:26.085 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.022) 0:00:26.108 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.024) 0:00:26.133 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587427.0269852, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587425.7599897, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 710, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587425.7599897, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.210) 0:00:26.344 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.026) 0:00:26.370 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.020) 0:00:26.390 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.024) 0:00:26.415 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.019) 0:00:26.435 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.017) 0:00:26.453 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.021) 0:00:26.474 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:43:51 +0000 (0:00:00.019) 0:00:26.493 ********* changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: cryptsetup-2.4.3-2.fc36.x86_64" ] } TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:43:54 +0000 (0:00:03.038) 0:00:29.532 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:43:54 +0000 (0:00:00.018) 0:00:29.550 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.568 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.024) 0:00:29.592 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.610 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.628 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.019) 0:00:29.647 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.665 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.683 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.022) 0:00:29.705 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.021) 0:00:29.727 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.745 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.019) 0:00:29.764 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.039) 0:00:29.803 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.018) 0:00:29.822 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.840 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.018) 0:00:29.859 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.018) 0:00:29.878 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.895 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.019) 0:00:29.915 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.933 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.017) 0:00:29.951 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.265) 0:00:30.216 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.201) 0:00:30.417 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.025) 0:00:30.442 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:43:55 +0000 (0:00:00.019) 0:00:30.462 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.199) 0:00:30.661 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.021) 0:00:30.683 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.021) 0:00:30.705 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.021) 0:00:30.726 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.021) 0:00:30.747 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.765 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.782 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.800 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.019) 0:00:30.819 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.837 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.855 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.018) 0:00:30.873 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.891 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.908 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.019) 0:00:30.928 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.945 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.963 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.981 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:30.999 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:31.016 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.020) 0:00:31.037 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.018) 0:00:31.056 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.024) 0:00:31.080 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.034122", "end": "2023-06-01 02:43:56.727269", "rc": 0, "start": "2023-06-01 02:43:56.693147" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.236) 0:00:31.317 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.024) 0:00:31.342 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.025) 0:00:31.367 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.020) 0:00:31.388 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.020) 0:00:31.408 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.020) 0:00:31.428 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.021) 0:00:31.450 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.017) 0:00:31.468 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.015) 0:00:31.483 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume_size to 9g] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:60 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.027) 0:00:31.510 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:43:56 +0000 (0:00:00.038) 0:00:31.549 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.025) 0:00:31.574 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.334) 0:00:31.909 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.045) 0:00:31.954 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.016) 0:00:31.971 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.016) 0:00:31.988 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.038) 0:00:32.026 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.017) 0:00:32.044 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "9g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.021) 0:00:32.066 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.018) 0:00:32.085 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.019) 0:00:32.104 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.018) 0:00:32.122 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.018) 0:00:32.140 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.017) 0:00:32.158 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.029) 0:00:32.188 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:43:57 +0000 (0:00:00.015) 0:00:32.203 ********* changed: [sut] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:43:59 +0000 (0:00:01.942) 0:00:34.145 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:43:59 +0000 (0:00:00.018) 0:00:34.164 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:43:59 +0000 (0:00:00.015) 0:00:34.180 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:43:59 +0000 (0:00:00.022) 0:00:34.202 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:43:59 +0000 (0:00:00.021) 0:00:34.224 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:43:59 +0000 (0:00:00.019) 0:00:34.244 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:43:59 +0000 (0:00:00.016) 0:00:34.261 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:44:00 +0000 (0:00:00.604) 0:00:34.865 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:44:00 +0000 (0:00:00.229) 0:00:35.095 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:44:00 +0000 (0:00:00.024) 0:00:35.119 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:44:01 +0000 (0:00:00.594) 0:00:35.714 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:44:01 +0000 (0:00:00.206) 0:00:35.921 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:44:01 +0000 (0:00:00.017) 0:00:35.939 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:74 Thursday 01 June 2023 02:44:01 +0000 (0:00:00.607) 0:00:36.546 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:44:02 +0000 (0:00:00.034) 0:00:36.581 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:44:02 +0000 (0:00:00.024) 0:00:36.605 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:44:02 +0000 (0:00:00.018) 0:00:36.624 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test1", "size": "9G", "type": "lvm", "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "ldgszG-feN2-QvqV-idFl-en1N-ZwC8-lFZ2Mr" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:44:02 +0000 (0:00:00.209) 0:00:36.833 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:01.004862", "end": "2023-06-01 02:44:03.449215", "rc": 0, "start": "2023-06-01 02:44:02.444353" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:44:03 +0000 (0:00:01.205) 0:00:38.039 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004480", "end": "2023-06-01 02:44:04.650692", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:44:03.646212" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:44:04 +0000 (0:00:01.200) 0:00:39.239 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:44:04 +0000 (0:00:00.038) 0:00:39.278 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:44:04 +0000 (0:00:00.017) 0:00:39.295 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:44:04 +0000 (0:00:00.040) 0:00:39.335 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:44:04 +0000 (0:00:00.025) 0:00:39.360 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.207) 0:00:39.568 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.022) 0:00:39.590 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.022) 0:00:39.613 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.023) 0:00:39.636 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.022) 0:00:39.659 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.021) 0:00:39.681 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.018) 0:00:39.699 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.026) 0:00:39.725 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.035) 0:00:39.760 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:39.778 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:39.796 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:39.813 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:39.831 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:39.849 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:39.868 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.018) 0:00:39.886 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.016) 0:00:39.903 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.036) 0:00:39.940 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.039) 0:00:39.979 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:39.998 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.020) 0:00:40.018 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.038 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.057 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.076 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.096 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.115 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.037) 0:00:40.153 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.038) 0:00:40.191 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:40.209 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:40.227 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:40.245 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.016) 0:00:40.262 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.061) 0:00:40.323 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.021) 0:00:40.345 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.021) 0:00:40.367 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.036) 0:00:40.403 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.021) 0:00:40.424 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.022) 0:00:40.446 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:40.464 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.017) 0:00:40.482 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.018) 0:00:40.500 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.520 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:44:05 +0000 (0:00:00.019) 0:00:40.539 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.041) 0:00:40.581 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.040) 0:00:40.621 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:40.640 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:40.658 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:40.676 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:40.695 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.019) 0:00:40.714 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.019) 0:00:40.734 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.017) 0:00:40.751 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.017) 0:00:40.769 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.035) 0:00:40.804 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.024) 0:00:40.829 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.084) 0:00:40.913 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.021) 0:00:40.935 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2178069, "block_size": 4096, "block_total": 2300135, "block_used": 122066, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 589813, "inode_total": 589824, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 8921370624, "size_total": 9421352960, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2178069, "block_size": 4096, "block_total": 2300135, "block_used": 122066, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 589813, "inode_total": 589824, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 8921370624, "size_total": 9421352960, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.026) 0:00:40.962 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.020) 0:00:40.982 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.023) 0:00:41.006 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.021) 0:00:41.028 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:41.046 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:41.065 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.017) 0:00:41.083 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.024) 0:00:41.108 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:41.126 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.018) 0:00:41.144 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.017) 0:00:41.162 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.017) 0:00:41.180 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.035) 0:00:41.215 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.022) 0:00:41.238 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.021) 0:00:41.259 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.017) 0:00:41.277 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.016) 0:00:41.294 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.022) 0:00:41.316 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.023) 0:00:41.340 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587440.5029392, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587439.5029426, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 710, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587439.5029426, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:44:06 +0000 (0:00:00.207) 0:00:41.547 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.023) 0:00:41.571 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.018) 0:00:41.590 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.022) 0:00:41.612 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.044) 0:00:41.657 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.019) 0:00:41.677 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.021) 0:00:41.698 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:44:07 +0000 (0:00:00.017) 0:00:41.716 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:44:09 +0000 (0:00:02.340) 0:00:44.056 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.019) 0:00:44.075 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.094 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.025) 0:00:44.120 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.138 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.157 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.175 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.019) 0:00:44.195 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.213 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.022) 0:00:44.236 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.021) 0:00:44.258 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.276 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.294 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.019) 0:00:44.314 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.017) 0:00:44.331 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.350 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.368 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.386 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.405 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.020) 0:00:44.425 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.443 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:44:09 +0000 (0:00:00.018) 0:00:44.462 ********* ok: [sut] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.202) 0:00:44.664 ********* ok: [sut] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.200) 0:00:44.864 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "9663676416" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.024) 0:00:44.889 ********* ok: [sut] => { "storage_test_expected_size": "9663676416" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.019) 0:00:44.908 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.200) 0:00:45.109 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.021) 0:00:45.131 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.021) 0:00:45.152 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.021) 0:00:45.174 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.021) 0:00:45.195 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.213 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.231 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.248 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.266 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.019) 0:00:45.285 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.303 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.018) 0:00:45.321 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.339 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.019) 0:00:45.359 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.017) 0:00:45.377 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.019) 0:00:45.396 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.026) 0:00:45.423 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.018) 0:00:45.441 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.018) 0:00:45.460 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.018) 0:00:45.478 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 9663676416, "changed": false, "failed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.019) 0:00:45.498 ********* ok: [sut] => { "storage_test_expected_size": "9663676416" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.020) 0:00:45.518 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:44:10 +0000 (0:00:00.025) 0:00:45.544 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.039621", "end": "2023-06-01 02:44:11.196225", "rc": 0, "start": "2023-06-01 02:44:11.156604" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.241) 0:00:45.785 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.024) 0:00:45.809 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.046) 0:00:45.855 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.021) 0:00:45.877 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.020) 0:00:45.897 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.019) 0:00:45.917 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.019) 0:00:45.937 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.017) 0:00:45.954 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.016) 0:00:45.971 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 5g] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:77 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.018) 0:00:45.990 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.038) 0:00:46.029 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.026) 0:00:46.055 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.334) 0:00:46.389 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.047) 0:00:46.437 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.019) 0:00:46.456 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.019) 0:00:46.476 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.038) 0:00:46.515 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.018) 0:00:46.533 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:11 +0000 (0:00:00.022) 0:00:46.556 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.019) 0:00:46.575 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.018) 0:00:46.593 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.020) 0:00:46.614 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.019) 0:00:46.633 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.018) 0:00:46.652 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.031) 0:00:46.683 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:12 +0000 (0:00:00.015) 0:00:46.699 ********* changed: [sut] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:44:14 +0000 (0:00:01.900) 0:00:48.599 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.027) 0:00:48.627 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.025) 0:00:48.652 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.022) 0:00:48.675 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.020) 0:00:48.696 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.020) 0:00:48.716 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.018) 0:00:48.735 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.592) 0:00:49.327 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:44:14 +0000 (0:00:00.226) 0:00:49.553 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:44:15 +0000 (0:00:00.024) 0:00:49.578 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:44:15 +0000 (0:00:00.595) 0:00:50.173 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:44:15 +0000 (0:00:00.206) 0:00:50.380 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:44:15 +0000 (0:00:00.017) 0:00:50.397 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:91 Thursday 01 June 2023 02:44:16 +0000 (0:00:00.598) 0:00:50.995 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:44:16 +0000 (0:00:00.037) 0:00:51.033 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:44:16 +0000 (0:00:00.023) 0:00:51.057 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:44:16 +0000 (0:00:00.018) 0:00:51.076 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "ldgszG-feN2-QvqV-idFl-en1N-ZwC8-lFZ2Mr" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:44:16 +0000 (0:00:00.203) 0:00:51.279 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003237", "end": "2023-06-01 02:44:16.890386", "rc": 0, "start": "2023-06-01 02:44:16.887149" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:44:16 +0000 (0:00:00.198) 0:00:51.477 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004403", "end": "2023-06-01 02:44:18.091137", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:44:17.086734" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:44:18 +0000 (0:00:01.203) 0:00:52.681 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.038) 0:00:52.719 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.019) 0:00:52.739 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.041) 0:00:52.780 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.025) 0:00:52.806 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.201) 0:00:53.008 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.022) 0:00:53.030 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.024) 0:00:53.055 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.023) 0:00:53.079 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.021) 0:00:53.100 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.023) 0:00:53.123 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.021) 0:00:53.144 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.027) 0:00:53.172 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.036) 0:00:53.209 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.019) 0:00:53.228 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.021) 0:00:53.249 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.018) 0:00:53.268 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.018) 0:00:53.286 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.019) 0:00:53.305 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.018) 0:00:53.324 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.020) 0:00:53.344 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.017) 0:00:53.361 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.036) 0:00:53.397 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.037) 0:00:53.435 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.020) 0:00:53.455 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.017) 0:00:53.473 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.017) 0:00:53.491 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.019) 0:00:53.510 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.018) 0:00:53.529 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:44:18 +0000 (0:00:00.019) 0:00:53.549 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:53.567 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.037) 0:00:53.604 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.083) 0:00:53.688 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.018) 0:00:53.706 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:53.724 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.018) 0:00:53.743 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.019) 0:00:53.763 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.039) 0:00:53.802 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.021) 0:00:53.823 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.021) 0:00:53.845 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.036) 0:00:53.882 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.021) 0:00:53.903 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.021) 0:00:53.924 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:53.941 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:53.959 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.019) 0:00:53.978 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:53.996 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.016) 0:00:54.012 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.042) 0:00:54.055 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.044) 0:00:54.099 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.020) 0:00:54.119 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.019) 0:00:54.139 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.018) 0:00:54.157 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.019) 0:00:54.176 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:54.194 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:54.212 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:54.229 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.017) 0:00:54.246 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.036) 0:00:54.282 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.021) 0:00:54.304 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.087) 0:00:54.392 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.022) 0:00:54.415 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1199011, "block_size": 4096, "block_total": 1268648, "block_used": 69637, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4911149056, "size_total": 5196382208, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1199011, "block_size": 4096, "block_total": 1268648, "block_used": 69637, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4911149056, "size_total": 5196382208, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.026) 0:00:54.442 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.019) 0:00:54.461 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.022) 0:00:54.484 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.021) 0:00:54.506 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.018) 0:00:54.524 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:44:19 +0000 (0:00:00.018) 0:00:54.543 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.019) 0:00:54.562 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.023) 0:00:54.586 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.018) 0:00:54.604 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.018) 0:00:54.622 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.018) 0:00:54.640 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.017) 0:00:54.658 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.039) 0:00:54.697 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.021) 0:00:54.719 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.021) 0:00:54.741 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.018) 0:00:54.759 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.017) 0:00:54.777 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.024) 0:00:54.802 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.049) 0:00:54.852 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587454.9608898, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587453.9368935, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 710, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587453.9368935, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.209) 0:00:55.061 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.023) 0:00:55.085 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.018) 0:00:55.104 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.021) 0:00:55.126 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.019) 0:00:55.146 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.017) 0:00:55.164 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.021) 0:00:55.185 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:44:20 +0000 (0:00:00.018) 0:00:55.203 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:44:22 +0000 (0:00:02.343) 0:00:57.547 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.018) 0:00:57.565 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.018) 0:00:57.583 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.024) 0:00:57.608 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.626 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.019) 0:00:57.645 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.663 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.681 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.018) 0:00:57.699 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.022) 0:00:57.721 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.021) 0:00:57.743 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.019) 0:00:57.762 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.018) 0:00:57.780 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.798 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.016) 0:00:57.815 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.833 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.851 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.019) 0:00:57.870 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.888 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.905 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.017) 0:00:57.923 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.018) 0:00:57.942 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.200) 0:00:58.143 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.200) 0:00:58.343 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.025) 0:00:58.369 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:44:23 +0000 (0:00:00.019) 0:00:58.389 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.200) 0:00:58.590 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.022) 0:00:58.613 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.023) 0:00:58.636 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.023) 0:00:58.660 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.021) 0:00:58.682 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:58.700 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.020) 0:00:58.720 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.027) 0:00:58.748 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:58.767 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.017) 0:00:58.784 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.017) 0:00:58.802 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:58.820 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.017) 0:00:58.838 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.019) 0:00:58.858 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.017) 0:00:58.876 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:58.894 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.017) 0:00:58.912 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:58.930 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:58.949 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.020) 0:00:58.969 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.020) 0:00:58.990 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.019) 0:00:59.009 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.024) 0:00:59.034 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.033689", "end": "2023-06-01 02:44:24.681527", "rc": 0, "start": "2023-06-01 02:44:24.647838" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.236) 0:00:59.271 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.024) 0:00:59.295 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.024) 0:00:59.319 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.020) 0:00:59.340 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.020) 0:00:59.361 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.020) 0:00:59.382 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.045) 0:00:59.427 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:59.445 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.015) 0:00:59.461 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Test for correct handling of too-large volume size] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:94 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.018) 0:00:59.480 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml for sut TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:4 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.039) 0:00:59.519 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:24 +0000 (0:00:00.030) 0:00:59.550 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.026) 0:00:59.577 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.336) 0:00:59.914 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.046) 0:00:59.960 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.018) 0:00:59.979 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.017) 0:00:59.996 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.039) 0:01:00.035 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.019) 0:01:00.054 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "12884901888.0" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.027) 0:01:00.082 ********* ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.025) 0:01:00.108 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.018) 0:01:00.126 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.019) 0:01:00.145 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.017) 0:01:00.163 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.019) 0:01:00.183 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.031) 0:01:00.214 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:25 +0000 (0:00:00.016) 0:01:00.230 ********* fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: volume 'test1' cannot be resized to '12 GiB' TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:95 Thursday 01 June 2023 02:44:27 +0000 (0:00:01.490) 0:01:01.721 ********* fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "volume 'test1' cannot be resized to '12 GiB'", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'ext4', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '12884901888.0', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.034) 0:01:01.755 ********* TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:20 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.018) 0:01:01.773 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:25 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.021) 0:01:01.795 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:36 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.025) 0:01:01.820 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Try to create LVM with volume size equal disk's size, resize to 10737418240] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:112 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.022) 0:01:01.843 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.029) 0:01:01.873 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.025) 0:01:01.899 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.335) 0:01:02.234 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.046) 0:01:02.280 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.016) 0:01:02.297 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.017) 0:01:02.314 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.037) 0:01:02.351 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.018) 0:01:02.369 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "10737418240" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.024) 0:01:02.394 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.048) 0:01:02.442 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.020) 0:01:02.463 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.019) 0:01:02.483 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.020) 0:01:02.503 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.019) 0:01:02.523 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:27 +0000 (0:00:00.032) 0:01:02.556 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:28 +0000 (0:00:00.017) 0:01:02.573 ********* changed: [sut] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:44:29 +0000 (0:00:01.912) 0:01:04.485 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:29 +0000 (0:00:00.019) 0:01:04.505 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:44:29 +0000 (0:00:00.015) 0:01:04.521 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:44:29 +0000 (0:00:00.022) 0:01:04.543 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:44:30 +0000 (0:00:00.021) 0:01:04.565 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:44:30 +0000 (0:00:00.018) 0:01:04.584 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:44:30 +0000 (0:00:00.016) 0:01:04.601 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:44:30 +0000 (0:00:00.596) 0:01:05.197 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:44:30 +0000 (0:00:00.224) 0:01:05.421 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext4', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext4", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:44:30 +0000 (0:00:00.024) 0:01:05.446 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:44:31 +0000 (0:00:00.598) 0:01:06.045 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:44:31 +0000 (0:00:00.205) 0:01:06.250 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:44:31 +0000 (0:00:00.017) 0:01:06.267 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:127 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.607) 0:01:06.874 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.029) 0:01:06.904 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "10737418240", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.023) 0:01:06.927 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.017) 0:01:06.945 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext4", "label": "", "name": "/dev/mapper/foo-test1", "size": "10G", "type": "lvm", "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "ldgszG-feN2-QvqV-idFl-en1N-ZwC8-lFZ2Mr" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.208) 0:01:07.153 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003455", "end": "2023-06-01 02:44:32.766319", "rc": 0, "start": "2023-06-01 02:44:32.762864" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext4 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.200) 0:01:07.353 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003226", "end": "2023-06-01 02:44:32.965268", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:44:32.962042" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:44:32 +0000 (0:00:00.200) 0:01:07.554 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.038) 0:01:07.593 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.017) 0:01:07.611 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.041) 0:01:07.652 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.025) 0:01:07.677 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.202) 0:01:07.880 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.022) 0:01:07.902 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.022) 0:01:07.925 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.022) 0:01:07.948 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.021) 0:01:07.969 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.023) 0:01:07.993 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.011 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.029) 0:01:08.041 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.063) 0:01:08.104 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.019) 0:01:08.124 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.142 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.160 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.019) 0:01:08.180 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.198 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.216 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.235 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.253 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.038) 0:01:08.291 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.036) 0:01:08.328 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.346 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.365 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.019) 0:01:08.384 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.402 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.421 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.439 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.458 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.039) 0:01:08.497 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.038) 0:01:08.536 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:44:33 +0000 (0:00:00.018) 0:01:08.554 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.018) 0:01:08.572 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:08.592 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:08.610 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.040) 0:01:08.650 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.021) 0:01:08.672 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.022) 0:01:08.694 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.035) 0:01:08.730 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.021) 0:01:08.751 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.021) 0:01:08.772 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.020) 0:01:08.793 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:08.812 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:08.832 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:08.851 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.018) 0:01:08.870 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.041) 0:01:08.911 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.038) 0:01:08.950 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:08.970 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.018) 0:01:08.989 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.018) 0:01:09.008 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.026 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.044 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.061 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.016) 0:01:09.078 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.016) 0:01:09.094 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.037) 0:01:09.131 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.022) 0:01:09.154 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.108) 0:01:09.263 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.023) 0:01:09.287 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2421050, "block_size": 4096, "block_total": 2556119, "block_used": 135069, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 655349, "inode_total": 655360, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 9916620800, "size_total": 10469863424, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2421050, "block_size": 4096, "block_total": 2556119, "block_used": 135069, "device": "/dev/mapper/foo-test1", "fstype": "ext4", "inode_available": 655349, "inode_total": 655360, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 9916620800, "size_total": 10469863424, "uuid": "40ac1a3d-2b5b-4000-9ca7-6a210e30be5e" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.026) 0:01:09.313 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:09.332 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.023) 0:01:09.355 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.021) 0:01:09.377 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.395 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:09.415 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.433 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.022) 0:01:09.455 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.473 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.491 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.017) 0:01:09.509 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:44:34 +0000 (0:00:00.019) 0:01:09.528 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext4 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.037) 0:01:09.566 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.023) 0:01:09.589 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.022) 0:01:09.612 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.018) 0:01:09.630 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.017) 0:01:09.648 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.024) 0:01:09.672 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.024) 0:01:09.696 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587470.8318357, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587469.8408391, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 710, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587469.8408391, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.206) 0:01:09.903 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.025) 0:01:09.928 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.018) 0:01:09.946 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.021) 0:01:09.968 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.019) 0:01:09.988 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.018) 0:01:10.006 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.021) 0:01:10.028 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:44:35 +0000 (0:00:00.019) 0:01:10.047 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:44:37 +0000 (0:00:02.365) 0:01:12.412 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.018) 0:01:12.431 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.018) 0:01:12.449 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.024) 0:01:12.474 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.018) 0:01:12.493 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.018) 0:01:12.511 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.017) 0:01:12.529 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:44:37 +0000 (0:00:00.017) 0:01:12.547 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.565 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.023) 0:01:12.588 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.022) 0:01:12.611 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.629 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.647 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.664 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.682 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.700 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.019) 0:01:12.719 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.042) 0:01:12.762 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.019) 0:01:12.782 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.800 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.017) 0:01:12.817 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.023) 0:01:12.841 ********* ok: [sut] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.202) 0:01:13.043 ********* ok: [sut] => { "bytes": 10737418240, "changed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.200) 0:01:13.243 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "10737418240" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.025) 0:01:13.268 ********* ok: [sut] => { "storage_test_expected_size": "10737418240" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.019) 0:01:13.287 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.199) 0:01:13.487 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.022) 0:01:13.510 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.021) 0:01:13.531 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:44:38 +0000 (0:00:00.021) 0:01:13.553 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.020) 0:01:13.573 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.592 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.610 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.020) 0:01:13.630 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.649 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.667 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.685 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.704 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.722 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.020) 0:01:13.742 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.760 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.779 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.797 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:13.816 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.017) 0:01:13.834 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.019) 0:01:13.853 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 10737418240, "changed": false, "failed": false, "lvm": "10g", "parted": "10GiB", "size": "10 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.020) 0:01:13.874 ********* ok: [sut] => { "storage_test_expected_size": "10737418240" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.019) 0:01:13.893 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.026) 0:01:13.920 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.033940", "end": "2023-06-01 02:44:39.567331", "rc": 0, "start": "2023-06-01 02:44:39.533391" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.236) 0:01:14.156 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.024) 0:01:14.180 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.026) 0:01:14.206 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.023) 0:01:14.230 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.022) 0:01:14.253 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.023) 0:01:14.276 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.021) 0:01:14.298 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.018) 0:01:14.316 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.016) 0:01:14.333 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Test for correct handling of invalid size specification] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:130 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.017) 0:01:14.350 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml for sut TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:4 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.031) 0:01:14.382 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.027) 0:01:14.410 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:39 +0000 (0:00:00.027) 0:01:14.437 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.339) 0:01:14.776 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.047) 0:01:14.824 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.017) 0:01:14.842 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.017) 0:01:14.859 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.039) 0:01:14.899 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.049) 0:01:14.949 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "xyz GiB" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.026) 0:01:14.976 ********* ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.025) 0:01:15.001 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.020) 0:01:15.022 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.018) 0:01:15.041 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.019) 0:01:15.061 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.020) 0:01:15.082 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.031) 0:01:15.113 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:40 +0000 (0:00:00.016) 0:01:15.130 ********* fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: invalid size specification 'xyz GiB' in pool 'foo' TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:95 Thursday 01 June 2023 02:44:41 +0000 (0:00:01.402) 0:01:16.533 ********* fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "invalid size specification 'xyz GiB' in pool 'foo'", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': '', 'fs_type': 'ext4', 'mount_options': None, 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': 'xyz GiB', 'state': 'present', 'type': None, 'cached': None, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:41 +0000 (0:00:00.023) 0:01:16.556 ********* TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:20 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.016) 0:01:16.573 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:25 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.020) 0:01:16.593 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:36 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.028) 0:01:16.622 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Test for correct handling of invalid size specification] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:146 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.018) 0:01:16.640 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml for sut TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:4 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.032) 0:01:16.673 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.030) 0:01:16.703 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.029) 0:01:16.733 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.337) 0:01:17.070 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.045) 0:01:17.116 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.018) 0:01:17.135 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.019) 0:01:17.154 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.042) 0:01:17.196 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.020) 0:01:17.217 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "volumes": [ { "fs_type": "ext4", "mount_point": "/opt/test1", "name": "test1", "size": "none" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.030) 0:01:17.247 ********* ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.023) 0:01:17.271 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.019) 0:01:17.291 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.018) 0:01:17.309 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.019) 0:01:17.328 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.020) 0:01:17.349 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.036) 0:01:17.385 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:42 +0000 (0:00:00.017) 0:01:17.402 ********* fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: invalid size specification 'none' in pool 'foo' TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:95 Thursday 01 June 2023 02:44:44 +0000 (0:00:01.384) 0:01:18.787 ********* fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "invalid size specification 'none' in pool 'foo'", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': None, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': None, 'fs_label': '', 'fs_type': 'ext4', 'mount_options': None, 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': 'none', 'state': 'present', 'type': None, 'cached': None, 'cache_devices': [], 'cache_mode': None, 'cache_size': None, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None}], 'raid_chunk_size': None}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.023) 0:01:18.810 ********* TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:20 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.015) 0:01:18.826 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:25 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.020) 0:01:18.846 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-failed.yml:36 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.025) 0:01:18.872 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up] **************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:162 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.017) 0:01:18.890 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.038) 0:01:18.929 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.051) 0:01:18.980 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.338) 0:01:19.319 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.045) 0:01:19.364 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.016) 0:01:19.381 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.016) 0:01:19.397 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.037) 0:01:19.435 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.018) 0:01:19.453 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.021) 0:01:19.475 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.019) 0:01:19.494 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.018) 0:01:19.513 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.019) 0:01:19.532 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:44 +0000 (0:00:00.018) 0:01:19.550 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:45 +0000 (0:00:00.018) 0:01:19.568 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:45 +0000 (0:00:00.030) 0:01:19.599 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:45 +0000 (0:00:00.015) 0:01:19.615 ********* changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:44:47 +0000 (0:00:02.075) 0:01:21.690 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:47 +0000 (0:00:00.018) 0:01:21.708 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:44:47 +0000 (0:00:00.015) 0:01:21.724 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext4" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "ext4", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:44:47 +0000 (0:00:00.023) 0:01:21.748 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:44:47 +0000 (0:00:00.022) 0:01:21.770 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:44:47 +0000 (0:00:00.020) 0:01:21.790 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext4'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext4", "mount_info": { "fstype": "ext4", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:44:47 +0000 (0:00:00.217) 0:01:22.007 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:44:48 +0000 (0:00:00.602) 0:01:22.610 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:44:48 +0000 (0:00:00.018) 0:01:22.629 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:44:48 +0000 (0:00:00.017) 0:01:22.646 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:44:48 +0000 (0:00:00.588) 0:01:23.235 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:44:48 +0000 (0:00:00.206) 0:01:23.441 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:44:48 +0000 (0:00:00.026) 0:01:23.468 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:175 Thursday 01 June 2023 02:44:49 +0000 (0:00:00.607) 0:01:24.075 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:44:49 +0000 (0:00:00.033) 0:01:24.109 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext4", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:44:49 +0000 (0:00:00.023) 0:01:24.132 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:44:49 +0000 (0:00:00.017) 0:01:24.150 ********* ok: [sut] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:44:49 +0000 (0:00:00.206) 0:01:24.357 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003340", "end": "2023-06-01 02:44:49.970664", "rc": 0, "start": "2023-06-01 02:44:49.967324" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:44:49 +0000 (0:00:00.200) 0:01:24.557 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004491", "end": "2023-06-01 02:44:51.170145", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:44:50.165654" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:44:51 +0000 (0:00:01.201) 0:01:25.759 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.036) 0:01:25.796 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.016) 0:01:25.813 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.038) 0:01:25.851 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.024) 0:01:25.875 ********* TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.015) 0:01:25.891 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.021) 0:01:25.912 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.047) 0:01:25.959 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.024) 0:01:25.983 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.021) 0:01:26.004 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.020) 0:01:26.025 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.017) 0:01:26.043 ********* TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.015) 0:01:26.058 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.034) 0:01:26.093 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.019) 0:01:26.113 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.131 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.149 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.168 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.017) 0:01:26.185 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.204 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.017) 0:01:26.221 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.016) 0:01:26.238 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.037) 0:01:26.276 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.037) 0:01:26.314 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.332 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.351 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.370 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.017) 0:01:26.387 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.406 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.017) 0:01:26.423 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.017) 0:01:26.441 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.037) 0:01:26.478 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.038) 0:01:26.517 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.535 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:44:51 +0000 (0:00:00.018) 0:01:26.553 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.019) 0:01:26.573 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.016) 0:01:26.590 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.040) 0:01:26.630 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.020) 0:01:26.650 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.016) 0:01:26.667 ********* TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.015) 0:01:26.683 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.016) 0:01:26.700 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.038) 0:01:26.738 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.040) 0:01:26.778 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:26.796 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:26.814 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:26.831 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:26.850 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:26.867 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:26.886 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:26.904 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.016) 0:01:26.921 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.035) 0:01:26.957 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.022) 0:01:26.980 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.109) 0:01:27.089 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.021) 0:01:27.110 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.024) 0:01:27.135 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.020) 0:01:27.155 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.019) 0:01:27.175 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.022) 0:01:27.197 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:27.216 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:27.235 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.253 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.019) 0:01:27.272 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.290 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.308 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.325 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.343 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.035) 0:01:27.379 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.019) 0:01:27.398 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.021) 0:01:27.419 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.437 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:27.455 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.018) 0:01:27.473 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:44:52 +0000 (0:00:00.017) 0:01:27.491 ********* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.196) 0:01:27.688 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.019) 0:01:27.708 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.022) 0:01:27.731 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.015) 0:01:27.746 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.018) 0:01:27.765 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.017) 0:01:27.782 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.014) 0:01:27.797 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:44:53 +0000 (0:00:00.017) 0:01:27.814 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:44:55 +0000 (0:00:02.343) 0:01:30.157 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.018) 0:01:30.176 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.018) 0:01:30.194 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.016) 0:01:30.210 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.228 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.246 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.015) 0:01:30.262 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.014) 0:01:30.277 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.014) 0:01:30.291 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.023) 0:01:30.315 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.021) 0:01:30.336 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.354 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.019) 0:01:30.373 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.391 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.408 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.426 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.018) 0:01:30.444 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.018) 0:01:30.463 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.019) 0:01:30.482 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.500 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.018) 0:01:30.518 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.018) 0:01:30.536 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:44:55 +0000 (0:00:00.017) 0:01:30.554 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.572 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.019) 0:01:30.592 ********* ok: [sut] => { "storage_test_expected_size": "10737418240" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.043) 0:01:30.636 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.655 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.673 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.690 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.708 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.726 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.019) 0:01:30.746 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.764 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.782 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.800 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.818 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.836 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.019) 0:01:30.855 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.873 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.892 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.910 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.928 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:30.946 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.019) 0:01:30.965 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:30.983 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:31.001 ********* ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.020 ********* ok: [sut] => { "storage_test_expected_size": "10737418240" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.038 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:31.056 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.019) 0:01:31.076 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.094 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.112 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.017) 0:01:31.130 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.148 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.167 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.021) 0:01:31.188 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.019) 0:01:31.207 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.016) 0:01:31.224 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a LVM logical volume with for ext3 FS size 5g] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:180 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.018) 0:01:31.242 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.046) 0:01:31.288 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:44:56 +0000 (0:00:00.026) 0:01:31.315 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.338) 0:01:31.653 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.046) 0:01:31.700 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.016) 0:01:31.717 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.017) 0:01:31.734 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.036) 0:01:31.771 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.019) 0:01:31.790 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext3", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.022) 0:01:31.813 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.019) 0:01:31.832 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.018) 0:01:31.850 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.017) 0:01:31.868 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.017) 0:01:31.886 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.017) 0:01:31.904 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.031) 0:01:31.936 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:44:57 +0000 (0:00:00.015) 0:01:31.952 ********* changed: [sut] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:44:59 +0000 (0:00:02.019) 0:01:33.971 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:44:59 +0000 (0:00:00.018) 0:01:33.990 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:44:59 +0000 (0:00:00.016) 0:01:34.006 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:44:59 +0000 (0:00:00.022) 0:01:34.029 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:44:59 +0000 (0:00:00.020) 0:01:34.049 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:44:59 +0000 (0:00:00.021) 0:01:34.071 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:44:59 +0000 (0:00:00.018) 0:01:34.089 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:45:00 +0000 (0:00:00.602) 0:01:34.692 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:45:00 +0000 (0:00:00.233) 0:01:34.925 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:45:00 +0000 (0:00:00.025) 0:01:34.951 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:45:00 +0000 (0:00:00.598) 0:01:35.549 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:45:01 +0000 (0:00:00.212) 0:01:35.762 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:45:01 +0000 (0:00:00.017) 0:01:35.779 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:196 Thursday 01 June 2023 02:45:01 +0000 (0:00:00.609) 0:01:36.389 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:45:01 +0000 (0:00:00.036) 0:01:36.426 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:45:01 +0000 (0:00:00.023) 0:01:36.449 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:45:01 +0000 (0:00:00.018) 0:01:36.468 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext3", "label": "", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "xfLiD4-ZV1p-Ffwd-tcuD-3e0F-LA1d-31ZX8t" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.214) 0:01:36.682 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003561", "end": "2023-06-01 02:45:02.300415", "rc": 0, "start": "2023-06-01 02:45:02.296854" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.207) 0:01:36.890 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003551", "end": "2023-06-01 02:45:02.505494", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:45:02.501943" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.203) 0:01:37.094 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.038) 0:01:37.133 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.019) 0:01:37.152 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.042) 0:01:37.194 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.026) 0:01:37.221 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.207) 0:01:37.428 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.022) 0:01:37.451 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.023) 0:01:37.474 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.024) 0:01:37.498 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.021) 0:01:37.520 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:45:02 +0000 (0:00:00.024) 0:01:37.544 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.563 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.027) 0:01:37.590 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.035) 0:01:37.626 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.020) 0:01:37.647 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.665 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.684 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.703 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.019) 0:01:37.722 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.741 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.020) 0:01:37.762 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.781 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.037) 0:01:37.818 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.038) 0:01:37.857 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.043) 0:01:37.901 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.020) 0:01:37.921 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.939 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.017) 0:01:37.957 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.017) 0:01:37.975 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:37.993 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.019) 0:01:38.012 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.036) 0:01:38.049 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.039) 0:01:38.089 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.019) 0:01:38.108 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:38.126 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.017) 0:01:38.144 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.016) 0:01:38.161 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.040) 0:01:38.202 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.021) 0:01:38.223 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.021) 0:01:38.245 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.034) 0:01:38.279 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.023) 0:01:38.303 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.022) 0:01:38.325 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:38.343 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:38.362 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.017) 0:01:38.380 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.017) 0:01:38.397 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:38.416 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.041) 0:01:38.458 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.040) 0:01:38.499 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.019) 0:01:38.518 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:38.537 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:45:03 +0000 (0:00:00.018) 0:01:38.555 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:38.573 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:38.591 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.019) 0:01:38.611 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.017) 0:01:38.629 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:38.647 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.034) 0:01:38.682 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.032) 0:01:38.714 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.089) 0:01:38.804 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.022) 0:01:38.826 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1205649, "block_size": 4096, "block_total": 1271208, "block_used": 65559, "device": "/dev/mapper/foo-test1", "fstype": "ext3", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4938338304, "size_total": 5206867968, "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1205649, "block_size": 4096, "block_total": 1271208, "block_used": 65559, "device": "/dev/mapper/foo-test1", "fstype": "ext3", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4938338304, "size_total": 5206867968, "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.025) 0:01:38.852 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.019) 0:01:38.871 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.024) 0:01:38.896 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.021) 0:01:38.917 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.017) 0:01:38.935 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.017) 0:01:38.953 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.017) 0:01:38.971 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.024) 0:01:38.995 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.020) 0:01:39.015 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.048) 0:01:39.064 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.019) 0:01:39.083 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:39.102 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext3 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.037) 0:01:39.139 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.022) 0:01:39.162 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.024) 0:01:39.186 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:39.204 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:39.223 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.022) 0:01:39.245 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.024) 0:01:39.269 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587500.3477347, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587499.3197381, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 997, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587499.3197381, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.206) 0:01:39.476 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.024) 0:01:39.500 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.018) 0:01:39.519 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.021) 0:01:39.540 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:45:04 +0000 (0:00:00.019) 0:01:39.560 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:45:05 +0000 (0:00:00.019) 0:01:39.580 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:45:05 +0000 (0:00:00.021) 0:01:39.602 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:45:05 +0000 (0:00:00.018) 0:01:39.620 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:45:07 +0000 (0:00:02.327) 0:01:41.947 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.021) 0:01:41.968 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:41.988 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.025) 0:01:42.013 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.031 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:42.051 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.070 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.088 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.020) 0:01:42.108 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.024) 0:01:42.133 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.023) 0:01:42.156 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:42.175 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.194 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:42.213 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:42.233 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:42.252 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.271 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.290 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.308 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.018) 0:01:42.326 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.020) 0:01:42.346 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:45:07 +0000 (0:00:00.019) 0:01:42.366 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.203) 0:01:42.569 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.201) 0:01:42.771 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.025) 0:01:42.796 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.019) 0:01:42.816 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.199) 0:01:43.015 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.022) 0:01:43.038 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.021) 0:01:43.059 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.021) 0:01:43.081 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.023) 0:01:43.105 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.123 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.142 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.019) 0:01:43.161 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.020) 0:01:43.181 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.200 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.020) 0:01:43.220 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.020) 0:01:43.240 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.019) 0:01:43.260 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.278 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.297 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.315 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.020) 0:01:43.335 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.353 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.372 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.018) 0:01:43.390 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.020) 0:01:43.411 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.019) 0:01:43.431 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:45:08 +0000 (0:00:00.026) 0:01:43.457 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.037954", "end": "2023-06-01 02:45:09.110773", "rc": 0, "start": "2023-06-01 02:45:09.072819" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.242) 0:01:43.699 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.024) 0:01:43.723 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.024) 0:01:43.748 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.021) 0:01:43.769 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.020) 0:01:43.790 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.020) 0:01:43.810 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.020) 0:01:43.830 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.017) 0:01:43.848 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.016) 0:01:43.864 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 9g] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:199 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.020) 0:01:43.884 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.048) 0:01:43.933 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.027) 0:01:43.960 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.337) 0:01:44.297 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.045) 0:01:44.342 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.016) 0:01:44.359 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.016) 0:01:44.376 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.038) 0:01:44.414 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.019) 0:01:44.433 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext3", "mount_point": "/opt/test1", "name": "test1", "size": "9g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.021) 0:01:44.455 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.018) 0:01:44.474 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.018) 0:01:44.492 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.019) 0:01:44.512 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.018) 0:01:44.530 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:45:09 +0000 (0:00:00.018) 0:01:44.549 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:45:10 +0000 (0:00:00.031) 0:01:44.580 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:45:10 +0000 (0:00:00.016) 0:01:44.596 ********* changed: [sut] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:45:12 +0000 (0:00:02.423) 0:01:47.020 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:45:12 +0000 (0:00:00.018) 0:01:47.039 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:45:12 +0000 (0:00:00.015) 0:01:47.054 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:45:12 +0000 (0:00:00.022) 0:01:47.077 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:45:12 +0000 (0:00:00.021) 0:01:47.098 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:45:12 +0000 (0:00:00.022) 0:01:47.120 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:45:12 +0000 (0:00:00.018) 0:01:47.139 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:45:13 +0000 (0:00:00.601) 0:01:47.741 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:45:13 +0000 (0:00:00.226) 0:01:47.967 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:45:13 +0000 (0:00:00.024) 0:01:47.992 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.596) 0:01:48.588 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.206) 0:01:48.795 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.017) 0:01:48.812 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:213 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.596) 0:01:49.408 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.037) 0:01:49.446 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.052) 0:01:49.498 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:45:14 +0000 (0:00:00.019) 0:01:49.518 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext3", "label": "", "name": "/dev/mapper/foo-test1", "size": "9G", "type": "lvm", "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "xfLiD4-ZV1p-Ffwd-tcuD-3e0F-LA1d-31ZX8t" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.208) 0:01:49.727 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003367", "end": "2023-06-01 02:45:15.339526", "rc": 0, "start": "2023-06-01 02:45:15.336159" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.200) 0:01:49.927 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003543", "end": "2023-06-01 02:45:15.541358", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:45:15.537815" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.202) 0:01:50.129 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.039) 0:01:50.169 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.017) 0:01:50.186 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.041) 0:01:50.228 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.026) 0:01:50.255 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.202) 0:01:50.457 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.021) 0:01:50.479 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.023) 0:01:50.502 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.022) 0:01:50.525 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:45:15 +0000 (0:00:00.021) 0:01:50.546 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.021) 0:01:50.567 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:50.585 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.027) 0:01:50.613 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.036) 0:01:50.649 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.019) 0:01:50.668 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.686 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.705 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.019) 0:01:50.725 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.743 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.762 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.780 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.799 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.037) 0:01:50.836 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.037) 0:01:50.873 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.891 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:50.909 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:50.928 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:50.946 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:50.963 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:50.981 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:50.999 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.037) 0:01:51.037 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.036) 0:01:51.074 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:51.092 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:51.110 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.019) 0:01:51.129 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:51.148 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.040) 0:01:51.189 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.022) 0:01:51.212 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.049) 0:01:51.261 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.037) 0:01:51.298 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.021) 0:01:51.320 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.022) 0:01:51.342 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.019) 0:01:51.362 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.019) 0:01:51.381 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.019) 0:01:51.401 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:51.419 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:51.437 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.040) 0:01:51.477 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.038) 0:01:51.515 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.018) 0:01:51.533 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:45:16 +0000 (0:00:00.017) 0:01:51.551 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:51.571 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:51.590 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.018) 0:01:51.608 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.017) 0:01:51.626 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.016) 0:01:51.643 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.017) 0:01:51.661 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.036) 0:01:51.698 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.022) 0:01:51.720 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.085) 0:01:51.806 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.023) 0:01:51.830 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2185028, "block_size": 4096, "block_total": 2303015, "block_used": 117987, "device": "/dev/mapper/foo-test1", "fstype": "ext3", "inode_available": 589813, "inode_total": 589824, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 8949874688, "size_total": 9433149440, "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2185028, "block_size": 4096, "block_total": 2303015, "block_used": 117987, "device": "/dev/mapper/foo-test1", "fstype": "ext3", "inode_available": 589813, "inode_total": 589824, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 8949874688, "size_total": 9433149440, "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.027) 0:01:51.857 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:51.877 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.022) 0:01:51.899 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.022) 0:01:51.922 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:51.941 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.020) 0:01:51.961 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.018) 0:01:51.979 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.022) 0:01:52.002 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.017) 0:01:52.020 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.017) 0:01:52.038 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.017) 0:01:52.056 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.020) 0:01:52.076 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext3 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.037) 0:01:52.113 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.022) 0:01:52.135 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.021) 0:01:52.157 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.020) 0:01:52.177 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:52.196 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.022) 0:01:52.219 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.023) 0:01:52.242 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587513.3756897, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587512.3746932, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 997, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587512.3746932, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.207) 0:01:52.449 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.025) 0:01:52.475 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:52.494 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.021) 0:01:52.516 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.019) 0:01:52.535 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:45:17 +0000 (0:00:00.017) 0:01:52.553 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:45:18 +0000 (0:00:00.021) 0:01:52.574 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:45:18 +0000 (0:00:00.045) 0:01:52.619 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:45:20 +0000 (0:00:02.343) 0:01:54.963 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:54.981 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:54.999 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.024) 0:01:55.024 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.019) 0:01:55.044 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.062 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.017) 0:01:55.080 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.017) 0:01:55.098 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.017) 0:01:55.115 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.024) 0:01:55.140 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.024) 0:01:55.165 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.019) 0:01:55.185 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.203 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.222 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.019) 0:01:55.241 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.260 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.020) 0:01:55.280 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.299 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.317 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.336 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.018) 0:01:55.354 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:45:20 +0000 (0:00:00.017) 0:01:55.372 ********* ok: [sut] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.203) 0:01:55.575 ********* ok: [sut] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.205) 0:01:55.781 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "9663676416" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.024) 0:01:55.805 ********* ok: [sut] => { "storage_test_expected_size": "9663676416" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.019) 0:01:55.825 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.199) 0:01:56.024 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.022) 0:01:56.047 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.021) 0:01:56.069 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.021) 0:01:56.091 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.021) 0:01:56.112 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.017) 0:01:56.130 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.017) 0:01:56.148 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.019) 0:01:56.168 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.018) 0:01:56.186 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.017) 0:01:56.204 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.017) 0:01:56.222 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.018) 0:01:56.240 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.017) 0:01:56.257 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.019) 0:01:56.277 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.018) 0:01:56.295 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.018) 0:01:56.314 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.020) 0:01:56.334 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.018) 0:01:56.353 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.022) 0:01:56.376 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.021) 0:01:56.397 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 9663676416, "changed": false, "failed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.021) 0:01:56.418 ********* ok: [sut] => { "storage_test_expected_size": "9663676416" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.019) 0:01:56.438 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:45:21 +0000 (0:00:00.025) 0:01:56.464 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.033903", "end": "2023-06-01 02:45:22.113645", "rc": 0, "start": "2023-06-01 02:45:22.079742" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.238) 0:01:56.702 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.024) 0:01:56.726 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.023) 0:01:56.750 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.020) 0:01:56.770 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.019) 0:01:56.790 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.021) 0:01:56.811 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.021) 0:01:56.833 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.017) 0:01:56.851 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.016) 0:01:56.867 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 5g] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:216 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.018) 0:01:56.885 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.053) 0:01:56.938 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.027) 0:01:56.965 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.333) 0:01:57.299 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.045) 0:01:57.344 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.017) 0:01:57.362 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.016) 0:01:57.378 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.037) 0:01:57.416 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.018) 0:01:57.434 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext3", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.021) 0:01:57.456 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.019) 0:01:57.475 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.018) 0:01:57.494 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.018) 0:01:57.513 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.018) 0:01:57.531 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:45:22 +0000 (0:00:00.020) 0:01:57.551 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:45:23 +0000 (0:00:00.030) 0:01:57.582 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:45:23 +0000 (0:00:00.016) 0:01:57.599 ********* changed: [sut] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:45:25 +0000 (0:00:02.791) 0:02:00.390 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:45:25 +0000 (0:00:00.020) 0:02:00.410 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:45:25 +0000 (0:00:00.017) 0:02:00.428 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:45:25 +0000 (0:00:00.022) 0:02:00.451 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:45:25 +0000 (0:00:00.021) 0:02:00.472 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:45:25 +0000 (0:00:00.019) 0:02:00.492 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:45:25 +0000 (0:00:00.016) 0:02:00.509 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:45:26 +0000 (0:00:00.598) 0:02:01.107 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:45:26 +0000 (0:00:00.229) 0:02:01.337 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext3', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext3", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:45:26 +0000 (0:00:00.025) 0:02:01.362 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:45:27 +0000 (0:00:00.594) 0:02:01.956 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:45:27 +0000 (0:00:00.205) 0:02:02.162 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:45:27 +0000 (0:00:00.017) 0:02:02.179 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:230 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.600) 0:02:02.779 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.039) 0:02:02.819 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.022) 0:02:02.842 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.017) 0:02:02.860 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext3", "label": "", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "xfLiD4-ZV1p-Ffwd-tcuD-3e0F-LA1d-31ZX8t" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.202) 0:02:03.062 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003244", "end": "2023-06-01 02:45:28.673220", "rc": 0, "start": "2023-06-01 02:45:28.669976" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext3 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.199) 0:02:03.262 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003338", "end": "2023-06-01 02:45:28.873804", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:45:28.870466" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.198) 0:02:03.460 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.038) 0:02:03.499 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:45:28 +0000 (0:00:00.017) 0:02:03.516 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.064) 0:02:03.580 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.025) 0:02:03.606 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.199) 0:02:03.805 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.021) 0:02:03.827 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.022) 0:02:03.850 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.023) 0:02:03.873 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.021) 0:02:03.895 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.021) 0:02:03.916 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.019) 0:02:03.936 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.027) 0:02:03.963 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.036) 0:02:04.000 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.018 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.037 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.055 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.019) 0:02:04.074 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.017) 0:02:04.092 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.110 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.024) 0:02:04.135 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.026) 0:02:04.161 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.038) 0:02:04.200 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.039) 0:02:04.239 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.019) 0:02:04.258 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.277 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.295 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.313 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.332 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.019) 0:02:04.351 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.020) 0:02:04.371 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.037) 0:02:04.409 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.037) 0:02:04.447 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.019) 0:02:04.467 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.485 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.504 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:45:29 +0000 (0:00:00.018) 0:02:04.522 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.039) 0:02:04.562 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.021) 0:02:04.583 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.021) 0:02:04.605 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.035) 0:02:04.641 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.023) 0:02:04.664 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.023) 0:02:04.688 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.019) 0:02:04.707 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:04.726 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.017) 0:02:04.744 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:04.762 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.017) 0:02:04.779 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.042) 0:02:04.822 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.038) 0:02:04.860 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:04.879 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.048) 0:02:04.928 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.020) 0:02:04.948 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:04.967 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.019) 0:02:04.987 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.020) 0:02:05.007 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:05.026 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.019) 0:02:05.045 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.036) 0:02:05.081 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.022) 0:02:05.104 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.089) 0:02:05.193 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.022) 0:02:05.216 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1205650, "block_size": 4096, "block_total": 1271208, "block_used": 65558, "device": "/dev/mapper/foo-test1", "fstype": "ext3", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4938342400, "size_total": 5206867968, "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1205650, "block_size": 4096, "block_total": 1271208, "block_used": 65558, "device": "/dev/mapper/foo-test1", "fstype": "ext3", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 4938342400, "size_total": 5206867968, "uuid": "7bc03eb3-841f-4575-9a00-d0042bb0a9cc" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.026) 0:02:05.243 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.019) 0:02:05.263 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.022) 0:02:05.286 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.021) 0:02:05.308 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.020) 0:02:05.328 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.019) 0:02:05.348 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:05.366 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.023) 0:02:05.390 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:05.409 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:05.427 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.020) 0:02:05.447 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.018) 0:02:05.465 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext3 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.036) 0:02:05.502 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.021) 0:02:05.524 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:45:30 +0000 (0:00:00.021) 0:02:05.545 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.019) 0:02:05.564 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.017) 0:02:05.582 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.022) 0:02:05.605 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.023) 0:02:05.628 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587526.7426436, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587525.7256472, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 997, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587525.7256472, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.208) 0:02:05.836 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.024) 0:02:05.860 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.018) 0:02:05.879 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.021) 0:02:05.900 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.018) 0:02:05.919 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.017) 0:02:05.937 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.022) 0:02:05.960 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:45:31 +0000 (0:00:00.018) 0:02:05.978 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:45:33 +0000 (0:00:02.328) 0:02:08.307 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.019) 0:02:08.326 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.018) 0:02:08.345 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.026) 0:02:08.371 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.018) 0:02:08.389 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.407 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.425 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.442 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.460 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.024) 0:02:08.485 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.021) 0:02:08.506 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.524 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.542 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:45:33 +0000 (0:00:00.017) 0:02:08.559 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:08.577 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.019) 0:02:08.596 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:08.614 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:08.632 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:08.650 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:08.668 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:08.686 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.051) 0:02:08.737 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.201) 0:02:08.938 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.202) 0:02:09.141 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.025) 0:02:09.166 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.019) 0:02:09.185 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.201) 0:02:09.386 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.021) 0:02:09.408 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.021) 0:02:09.429 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.021) 0:02:09.450 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.021) 0:02:09.472 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.018) 0:02:09.490 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.019) 0:02:09.510 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.018) 0:02:09.528 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:45:34 +0000 (0:00:00.017) 0:02:09.546 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:09.564 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.017) 0:02:09.582 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:09.600 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.020) 0:02:09.620 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:09.639 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:09.657 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:09.675 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.017) 0:02:09.693 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.017) 0:02:09.711 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.019) 0:02:09.730 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:09.748 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.020) 0:02:09.769 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.019) 0:02:09.788 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.025) 0:02:09.814 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.033996", "end": "2023-06-01 02:45:35.461542", "rc": 0, "start": "2023-06-01 02:45:35.427546" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.236) 0:02:10.050 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.024) 0:02:10.075 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.024) 0:02:10.099 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.020) 0:02:10.120 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.021) 0:02:10.141 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.020) 0:02:10.162 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.020) 0:02:10.182 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.018) 0:02:10.200 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.015) 0:02:10.216 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:233 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.016) 0:02:10.232 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.054) 0:02:10.287 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:45:35 +0000 (0:00:00.027) 0:02:10.314 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.335) 0:02:10.650 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.045) 0:02:10.696 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.016) 0:02:10.713 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.016) 0:02:10.729 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.037) 0:02:10.767 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.017) 0:02:10.785 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.021) 0:02:10.807 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.018) 0:02:10.826 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.017) 0:02:10.844 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.020) 0:02:10.864 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.019) 0:02:10.883 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.018) 0:02:10.902 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.030) 0:02:10.932 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:45:36 +0000 (0:00:00.015) 0:02:10.948 ********* changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "ext3", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:45:38 +0000 (0:00:02.091) 0:02:13.039 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:45:38 +0000 (0:00:00.018) 0:02:13.057 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:45:38 +0000 (0:00:00.015) 0:02:13.073 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext3" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "ext3", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:45:38 +0000 (0:00:00.022) 0:02:13.095 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:45:38 +0000 (0:00:00.022) 0:02:13.118 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:45:38 +0000 (0:00:00.020) 0:02:13.138 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext3'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext3", "mount_info": { "fstype": "ext3", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:45:38 +0000 (0:00:00.217) 0:02:13.356 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:45:39 +0000 (0:00:00.595) 0:02:13.952 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:45:39 +0000 (0:00:00.018) 0:02:13.971 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:45:39 +0000 (0:00:00.018) 0:02:13.989 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.591) 0:02:14.580 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.203) 0:02:14.784 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.016) 0:02:14.801 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:246 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.597) 0:02:15.399 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.043) 0:02:15.443 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext3", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.022) 0:02:15.465 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:45:40 +0000 (0:00:00.018) 0:02:15.484 ********* ok: [sut] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.207) 0:02:15.691 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003398", "end": "2023-06-01 02:45:41.302656", "rc": 0, "start": "2023-06-01 02:45:41.299258" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.198) 0:02:15.889 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003476", "end": "2023-06-01 02:45:41.504584", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:45:41.501108" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.203) 0:02:16.093 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.038) 0:02:16.131 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.018) 0:02:16.150 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.040) 0:02:16.190 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.025) 0:02:16.215 ********* TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.015) 0:02:16.231 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.020) 0:02:16.252 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.022) 0:02:16.274 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.025) 0:02:16.300 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.022) 0:02:16.322 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.021) 0:02:16.344 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.018) 0:02:16.363 ********* TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.015) 0:02:16.378 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.034) 0:02:16.413 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.017) 0:02:16.430 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.017) 0:02:16.448 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.017) 0:02:16.466 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.017) 0:02:16.483 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.018) 0:02:16.502 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:45:41 +0000 (0:00:00.042) 0:02:16.544 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.019) 0:02:16.564 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:16.582 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.034) 0:02:16.617 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.037) 0:02:16.655 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.673 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.690 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.707 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.725 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.019) 0:02:16.744 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.762 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.779 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.035) 0:02:16.814 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.036) 0:02:16.851 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.869 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.887 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.904 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:16.922 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.041) 0:02:16.963 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.020) 0:02:16.984 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.015) 0:02:17.000 ********* TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.016) 0:02:17.016 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:17.035 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.039) 0:02:17.075 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.037) 0:02:17.113 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.019) 0:02:17.132 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:17.150 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.019) 0:02:17.169 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:17.188 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:17.206 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:17.223 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:17.242 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.016) 0:02:17.258 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.033) 0:02:17.292 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.022) 0:02:17.314 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.085) 0:02:17.400 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.022) 0:02:17.422 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.025) 0:02:17.447 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.018) 0:02:17.466 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:17.484 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.022) 0:02:17.506 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:17.523 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:17.541 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:45:42 +0000 (0:00:00.017) 0:02:17.558 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.018) 0:02:17.576 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:17.594 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.019) 0:02:17.614 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.045) 0:02:17.660 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.018) 0:02:17.678 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.036) 0:02:17.715 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:17.732 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.021) 0:02:17.753 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.018) 0:02:17.772 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:17.790 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:17.807 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:17.825 ********* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.198) 0:02:18.023 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.018) 0:02:18.042 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.022) 0:02:18.064 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.015) 0:02:18.079 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.020) 0:02:18.100 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:18.117 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.015) 0:02:18.133 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:45:43 +0000 (0:00:00.017) 0:02:18.151 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:45:45 +0000 (0:00:02.337) 0:02:20.488 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:45:45 +0000 (0:00:00.019) 0:02:20.508 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:45:45 +0000 (0:00:00.017) 0:02:20.526 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:45:45 +0000 (0:00:00.014) 0:02:20.541 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:45:45 +0000 (0:00:00.017) 0:02:20.558 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.577 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.015) 0:02:20.592 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.016) 0:02:20.608 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.015) 0:02:20.624 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.024) 0:02:20.649 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.021) 0:02:20.670 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.689 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.708 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.726 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.743 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:20.762 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.780 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.799 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.816 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.834 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.852 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:20.872 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.890 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.908 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:20.925 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:20.945 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:20.964 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:20.984 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.001 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.019 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.038 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.055 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.074 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:21.094 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.112 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.130 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.148 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.166 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.026) 0:02:21.193 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.022) 0:02:21.215 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.234 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.251 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.269 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.288 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.306 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:21.325 ********* ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.020) 0:02:21.346 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.364 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.383 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.401 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.418 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:21.437 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.018) 0:02:21.456 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.474 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.491 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.509 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.017) 0:02:21.526 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:45:46 +0000 (0:00:00.019) 0:02:21.546 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a LVM logical volume with for ext2 FS size 5g] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:251 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.020) 0:02:21.566 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.089) 0:02:21.656 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.026) 0:02:21.682 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.342) 0:02:22.025 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.046) 0:02:22.071 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.018) 0:02:22.089 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.018) 0:02:22.108 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.038) 0:02:22.146 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.019) 0:02:22.166 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext2", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.022) 0:02:22.189 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.020) 0:02:22.209 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.019) 0:02:22.228 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.019) 0:02:22.248 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.019) 0:02:22.268 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.019) 0:02:22.287 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.030) 0:02:22.318 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:45:47 +0000 (0:00:00.018) 0:02:22.336 ********* changed: [sut] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:45:49 +0000 (0:00:01.903) 0:02:24.240 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:45:49 +0000 (0:00:00.018) 0:02:24.259 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:45:49 +0000 (0:00:00.015) 0:02:24.275 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/foo-test1" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:45:49 +0000 (0:00:00.023) 0:02:24.298 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:45:49 +0000 (0:00:00.021) 0:02:24.319 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:45:49 +0000 (0:00:00.020) 0:02:24.340 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:45:49 +0000 (0:00:00.018) 0:02:24.359 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:45:50 +0000 (0:00:00.604) 0:02:24.963 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:45:50 +0000 (0:00:00.349) 0:02:25.312 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:45:50 +0000 (0:00:00.024) 0:02:25.337 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:45:51 +0000 (0:00:00.880) 0:02:26.218 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:45:51 +0000 (0:00:00.239) 0:02:26.458 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:45:51 +0000 (0:00:00.017) 0:02:26.475 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:267 Thursday 01 June 2023 02:45:52 +0000 (0:00:00.870) 0:02:27.345 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:45:52 +0000 (0:00:00.048) 0:02:27.394 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:45:52 +0000 (0:00:00.025) 0:02:27.419 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:45:52 +0000 (0:00:00.019) 0:02:27.438 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext2", "label": "", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "0871M6-Ngap-iHST-j18Y-YRT3-9UsS-lYSsQZ" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.309) 0:02:27.748 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.004600", "end": "2023-06-01 02:45:53.468717", "rc": 0, "start": "2023-06-01 02:45:53.464117" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.310) 0:02:28.058 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.004563", "end": "2023-06-01 02:45:53.717821", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:45:53.713258" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.253) 0:02:28.312 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.038) 0:02:28.351 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.018) 0:02:28.369 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.040) 0:02:28.410 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:45:53 +0000 (0:00:00.025) 0:02:28.435 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.296) 0:02:28.731 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.022) 0:02:28.754 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.023) 0:02:28.777 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.023) 0:02:28.801 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.022) 0:02:28.823 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.021) 0:02:28.845 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:28.864 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.027) 0:02:28.891 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.038) 0:02:28.930 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.019) 0:02:28.949 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.019) 0:02:28.968 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:28.987 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.006 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.024 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.020) 0:02:29.045 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.064 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.017) 0:02:29.082 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.037) 0:02:29.119 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.039) 0:02:29.158 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.177 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.196 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.214 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.232 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.020) 0:02:29.253 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.019) 0:02:29.272 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.291 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.037) 0:02:29.328 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.039) 0:02:29.368 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.386 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.404 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.423 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.018) 0:02:29.441 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.040) 0:02:29.482 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.021) 0:02:29.503 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:45:54 +0000 (0:00:00.021) 0:02:29.525 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.036) 0:02:29.562 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.049) 0:02:29.612 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.023) 0:02:29.635 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.653 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.672 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.691 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.709 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.019) 0:02:29.729 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.042) 0:02:29.771 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.040) 0:02:29.812 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.830 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.019) 0:02:29.849 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.867 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.886 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.905 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.020) 0:02:29.925 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:29.943 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.016) 0:02:29.960 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.035) 0:02:29.995 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.023) 0:02:30.019 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.084) 0:02:30.103 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.021) 0:02:30.125 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1222050, "block_size": 4096, "block_total": 1287592, "block_used": 65542, "device": "/dev/mapper/foo-test1", "fstype": "ext2", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 5005516800, "size_total": 5273976832, "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1222050, "block_size": 4096, "block_total": 1287592, "block_used": 65542, "device": "/dev/mapper/foo-test1", "fstype": "ext2", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 5005516800, "size_total": 5273976832, "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.026) 0:02:30.151 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.019) 0:02:30.171 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.023) 0:02:30.195 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.022) 0:02:30.217 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.019) 0:02:30.236 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.020) 0:02:30.256 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.019) 0:02:30.275 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.023) 0:02:30.298 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.020) 0:02:30.319 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:30.337 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:30.356 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:30.374 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext2 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.037) 0:02:30.412 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.022) 0:02:30.435 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.021) 0:02:30.456 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.018) 0:02:30.475 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.019) 0:02:30.495 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.022) 0:02:30.517 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:45:55 +0000 (0:00:00.023) 0:02:30.541 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587550.7375607, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587549.5905647, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1198, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587549.5905647, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.207) 0:02:30.748 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.024) 0:02:30.773 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.018) 0:02:30.792 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.022) 0:02:30.814 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.020) 0:02:30.834 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.018) 0:02:30.853 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.022) 0:02:30.875 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:45:56 +0000 (0:00:00.018) 0:02:30.894 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:45:58 +0000 (0:00:02.325) 0:02:33.219 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.021) 0:02:33.240 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.051) 0:02:33.291 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.026) 0:02:33.318 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.336 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.354 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.373 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.391 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.019) 0:02:33.411 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.023) 0:02:33.434 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.021) 0:02:33.456 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.019) 0:02:33.475 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.494 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.512 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.019) 0:02:33.532 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:45:58 +0000 (0:00:00.018) 0:02:33.550 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:33.569 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:33.587 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:33.605 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:33.624 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.019) 0:02:33.644 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.017) 0:02:33.662 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.202) 0:02:33.864 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.202) 0:02:34.066 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.026) 0:02:34.093 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.021) 0:02:34.115 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.203) 0:02:34.319 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.024) 0:02:34.343 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.022) 0:02:34.365 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.022) 0:02:34.387 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.023) 0:02:34.411 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:34.430 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:34.449 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:34.467 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.018) 0:02:34.485 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.019) 0:02:34.505 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.023) 0:02:34.528 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:45:59 +0000 (0:00:00.021) 0:02:34.549 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.019) 0:02:34.568 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:34.587 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:34.606 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.019) 0:02:34.625 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.020) 0:02:34.646 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:34.664 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:34.683 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:34.701 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.019) 0:02:34.721 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.019) 0:02:34.741 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.028) 0:02:34.770 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.038672", "end": "2023-06-01 02:46:00.425853", "rc": 0, "start": "2023-06-01 02:46:00.387181" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.245) 0:02:35.015 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.026) 0:02:35.041 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.027) 0:02:35.069 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.021) 0:02:35.090 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.021) 0:02:35.112 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.021) 0:02:35.133 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.021) 0:02:35.154 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:35.173 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.018) 0:02:35.191 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 9g] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:270 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.019) 0:02:35.211 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.108) 0:02:35.319 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:46:00 +0000 (0:00:00.029) 0:02:35.348 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.339) 0:02:35.688 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.048) 0:02:35.736 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.017) 0:02:35.754 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.019) 0:02:35.773 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.040) 0:02:35.814 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.020) 0:02:35.835 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext2", "mount_point": "/opt/test1", "name": "test1", "size": "9g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.024) 0:02:35.859 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.022) 0:02:35.882 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.021) 0:02:35.904 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.020) 0:02:35.924 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.020) 0:02:35.945 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.020) 0:02:35.965 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.032) 0:02:35.998 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:46:01 +0000 (0:00:00.019) 0:02:36.017 ********* changed: [sut] => { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:46:03 +0000 (0:00:02.418) 0:02:38.436 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:46:03 +0000 (0:00:00.020) 0:02:38.456 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:46:03 +0000 (0:00:00.016) 0:02:38.472 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:46:03 +0000 (0:00:00.023) 0:02:38.496 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:46:03 +0000 (0:00:00.021) 0:02:38.518 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:46:03 +0000 (0:00:00.020) 0:02:38.538 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:46:03 +0000 (0:00:00.018) 0:02:38.556 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:46:04 +0000 (0:00:00.597) 0:02:39.153 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:46:04 +0000 (0:00:00.226) 0:02:39.380 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:46:04 +0000 (0:00:00.024) 0:02:39.404 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:46:05 +0000 (0:00:00.606) 0:02:40.011 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:46:05 +0000 (0:00:00.205) 0:02:40.217 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:46:05 +0000 (0:00:00.018) 0:02:40.235 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:284 Thursday 01 June 2023 02:46:06 +0000 (0:00:00.604) 0:02:40.839 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:46:06 +0000 (0:00:00.047) 0:02:40.887 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "9g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:46:06 +0000 (0:00:00.023) 0:02:40.910 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:46:06 +0000 (0:00:00.017) 0:02:40.928 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext2", "label": "", "name": "/dev/mapper/foo-test1", "size": "9G", "type": "lvm", "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "0871M6-Ngap-iHST-j18Y-YRT3-9UsS-lYSsQZ" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:46:06 +0000 (0:00:00.205) 0:02:41.133 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003425", "end": "2023-06-01 02:46:06.747350", "rc": 0, "start": "2023-06-01 02:46:06.743925" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:46:06 +0000 (0:00:00.202) 0:02:41.336 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004550", "end": "2023-06-01 02:46:07.951878", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:46:06.947328" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:46:07 +0000 (0:00:01.205) 0:02:42.542 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.069) 0:02:42.612 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:42.630 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.039) 0:02:42.669 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.025) 0:02:42.695 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.202) 0:02:42.897 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.021) 0:02:42.919 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.022) 0:02:42.942 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.024) 0:02:42.967 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.021) 0:02:42.988 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.021) 0:02:43.010 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.028 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.027) 0:02:43.056 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.036) 0:02:43.092 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.111 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.131 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.151 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.020) 0:02:43.171 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.020) 0:02:43.192 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.211 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.230 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.249 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.038) 0:02:43.287 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.037) 0:02:43.325 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.345 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.363 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.381 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.401 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.019) 0:02:43.421 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.439 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.017) 0:02:43.457 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.037) 0:02:43.495 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.037) 0:02:43.532 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:46:08 +0000 (0:00:00.018) 0:02:43.550 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:43.568 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:43.587 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.021) 0:02:43.608 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.042) 0:02:43.650 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.020) 0:02:43.671 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.021) 0:02:43.693 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.038) 0:02:43.731 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.022) 0:02:43.754 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.022) 0:02:43.776 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:43.795 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:43.813 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.019) 0:02:43.833 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:43.852 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.017) 0:02:43.869 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.041) 0:02:43.910 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.071) 0:02:43.982 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.022) 0:02:44.004 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.020) 0:02:44.025 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.020) 0:02:44.046 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.020) 0:02:44.066 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.022) 0:02:44.089 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.020) 0:02:44.109 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.128 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.146 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.038) 0:02:44.184 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.023) 0:02:44.207 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.087) 0:02:44.295 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.021) 0:02:44.317 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2201429, "block_size": 4096, "block_total": 2319399, "block_used": 117970, "device": "/dev/mapper/foo-test1", "fstype": "ext2", "inode_available": 589813, "inode_total": 589824, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 9017053184, "size_total": 9500258304, "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2201429, "block_size": 4096, "block_total": 2319399, "block_used": 117970, "device": "/dev/mapper/foo-test1", "fstype": "ext2", "inode_available": 589813, "inode_total": 589824, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 9017053184, "size_total": 9500258304, "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.026) 0:02:44.343 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.020) 0:02:44.364 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.023) 0:02:44.387 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.021) 0:02:44.409 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.428 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.447 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.465 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.024) 0:02:44.490 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.509 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.527 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:46:09 +0000 (0:00:00.018) 0:02:44.545 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.018) 0:02:44.564 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext2 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.038) 0:02:44.602 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.022) 0:02:44.625 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.021) 0:02:44.647 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.018) 0:02:44.665 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.017) 0:02:44.682 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.022) 0:02:44.705 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.025) 0:02:44.731 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587564.790512, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587563.7895155, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1198, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587563.7895155, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.209) 0:02:44.941 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.025) 0:02:44.966 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.019) 0:02:44.986 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.022) 0:02:45.008 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.019) 0:02:45.028 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.018) 0:02:45.047 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.020) 0:02:45.068 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:46:10 +0000 (0:00:00.018) 0:02:45.086 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:46:12 +0000 (0:00:02.365) 0:02:47.452 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:46:12 +0000 (0:00:00.020) 0:02:47.472 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:46:12 +0000 (0:00:00.018) 0:02:47.491 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:46:12 +0000 (0:00:00.025) 0:02:47.516 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:46:12 +0000 (0:00:00.018) 0:02:47.535 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:46:12 +0000 (0:00:00.018) 0:02:47.553 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.020) 0:02:47.574 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.020) 0:02:47.595 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.019) 0:02:47.615 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.026) 0:02:47.641 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.024) 0:02:47.665 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.023) 0:02:47.688 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.068) 0:02:47.757 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.021) 0:02:47.778 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.020) 0:02:47.799 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.020) 0:02:47.819 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.019) 0:02:47.839 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.019) 0:02:47.858 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.020) 0:02:47.879 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.018) 0:02:47.898 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.019) 0:02:47.917 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.019) 0:02:47.937 ********* ok: [sut] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.204) 0:02:48.141 ********* ok: [sut] => { "bytes": 9663676416, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.204) 0:02:48.345 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "9663676416" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.028) 0:02:48.374 ********* ok: [sut] => { "storage_test_expected_size": "9663676416" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:46:13 +0000 (0:00:00.023) 0:02:48.397 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.201) 0:02:48.598 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.022) 0:02:48.621 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:48.643 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.022) 0:02:48.666 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:48.687 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.018) 0:02:48.706 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.018) 0:02:48.724 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:48.745 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.019) 0:02:48.765 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.020) 0:02:48.786 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.018) 0:02:48.805 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:48.827 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.028) 0:02:48.855 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:48.876 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.019) 0:02:48.895 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.020) 0:02:48.916 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.020) 0:02:48.936 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.018) 0:02:48.955 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.018) 0:02:48.974 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.019) 0:02:48.993 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 9663676416, "changed": false, "failed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.020) 0:02:49.013 ********* ok: [sut] => { "storage_test_expected_size": "9663676416" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:49.035 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.026) 0:02:49.062 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.034493", "end": "2023-06-01 02:46:14.710703", "rc": 0, "start": "2023-06-01 02:46:14.676210" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.239) 0:02:49.301 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.025) 0:02:49.327 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.025) 0:02:49.352 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:49.373 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:49.395 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.020) 0:02:49.416 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.021) 0:02:49.437 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.020) 0:02:49.458 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.016) 0:02:49.474 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change volume size to 5g] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:287 Thursday 01 June 2023 02:46:14 +0000 (0:00:00.018) 0:02:49.493 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.072) 0:02:49.566 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.027) 0:02:49.593 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.337) 0:02:49.931 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.048) 0:02:49.980 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.018) 0:02:49.998 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.017) 0:02:50.015 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.037) 0:02:50.053 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.057) 0:02:50.111 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "fs_type": "ext2", "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.025) 0:02:50.137 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.020) 0:02:50.157 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.020) 0:02:50.178 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.020) 0:02:50.199 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.021) 0:02:50.220 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.022) 0:02:50.242 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.032) 0:02:50.275 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:46:15 +0000 (0:00:00.018) 0:02:50.293 ********* changed: [sut] => { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:46:18 +0000 (0:00:02.784) 0:02:53.077 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:46:18 +0000 (0:00:00.018) 0:02:53.096 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:46:18 +0000 (0:00:00.018) 0:02:53.114 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "resize format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "resize device", "device": "/dev/mapper/foo-test1", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:46:18 +0000 (0:00:00.023) 0:02:53.137 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:46:18 +0000 (0:00:00.022) 0:02:53.160 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:46:18 +0000 (0:00:00.021) 0:02:53.182 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:46:18 +0000 (0:00:00.019) 0:02:53.201 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:46:19 +0000 (0:00:00.596) 0:02:53.797 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:46:19 +0000 (0:00:00.226) 0:02:54.024 ********* skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'ext2', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "ext2", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:46:19 +0000 (0:00:00.026) 0:02:54.051 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:46:20 +0000 (0:00:00.593) 0:02:54.644 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:46:20 +0000 (0:00:00.205) 0:02:54.849 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:46:20 +0000 (0:00:00.018) 0:02:54.868 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:301 Thursday 01 June 2023 02:46:20 +0000 (0:00:00.610) 0:02:55.479 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:46:20 +0000 (0:00:00.053) 0:02:55.532 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:46:20 +0000 (0:00:00.023) 0:02:55.555 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:46:21 +0000 (0:00:00.017) 0:02:55.573 ********* ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "ext2", "label": "", "name": "/dev/mapper/foo-test1", "size": "5G", "type": "lvm", "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "0871M6-Ngap-iHST-j18Y-YRT3-9UsS-lYSsQZ" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:46:21 +0000 (0:00:00.205) 0:02:55.779 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003284", "end": "2023-06-01 02:46:21.390491", "rc": 0, "start": "2023-06-01 02:46:21.387207" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 ext2 defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:46:21 +0000 (0:00:00.199) 0:02:55.979 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004527", "end": "2023-06-01 02:46:22.592123", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:46:21.587596" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:46:22 +0000 (0:00:01.202) 0:02:57.181 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:46:22 +0000 (0:00:00.070) 0:02:57.252 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:46:22 +0000 (0:00:00.019) 0:02:57.272 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:46:22 +0000 (0:00:00.041) 0:02:57.313 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:46:22 +0000 (0:00:00.026) 0:02:57.339 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:46:22 +0000 (0:00:00.202) 0:02:57.542 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.022) 0:02:57.564 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.022) 0:02:57.587 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.024) 0:02:57.611 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.021) 0:02:57.633 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.021) 0:02:57.655 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.673 ********* ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.026) 0:02:57.700 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.037) 0:02:57.737 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.755 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.017) 0:02:57.773 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.017) 0:02:57.791 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.017) 0:02:57.809 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.019) 0:02:57.829 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.847 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.866 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.884 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.037) 0:02:57.922 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.038) 0:02:57.961 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.980 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:57.998 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.017) 0:02:58.016 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.019) 0:02:58.035 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:58.054 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.020) 0:02:58.074 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:58.093 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.039) 0:02:58.132 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.038) 0:02:58.171 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:58.190 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:58.208 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:58.227 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.019) 0:02:58.246 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.042) 0:02:58.288 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.023) 0:02:58.312 ********* skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.022) 0:02:58.335 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:2 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.040) 0:02:58.375 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:9 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.024) 0:02:58.399 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:18 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.024) 0:02:58.424 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:27 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.019) 0:02:58.444 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:37 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.019) 0:02:58.463 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-crypttab.yml:47 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.020) 0:02:58.484 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.018) 0:02:58.502 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:46:23 +0000 (0:00:00.017) 0:02:58.520 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.040) 0:02:58.560 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.071) 0:02:58.631 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.021) 0:02:58.653 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.021) 0:02:58.675 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.020) 0:02:58.695 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.020) 0:02:58.716 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.020) 0:02:58.736 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.019) 0:02:58.756 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.018) 0:02:58.774 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.018) 0:02:58.792 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.043) 0:02:58.836 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.024) 0:02:58.860 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.089) 0:02:58.950 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.022) 0:02:58.973 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1222051, "block_size": 4096, "block_total": 1287592, "block_used": 65541, "device": "/dev/mapper/foo-test1", "fstype": "ext2", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 5005520896, "size_total": 5273976832, "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1222051, "block_size": 4096, "block_total": 1287592, "block_used": 65541, "device": "/dev/mapper/foo-test1", "fstype": "ext2", "inode_available": 327669, "inode_total": 327680, "inode_used": 11, "mount": "/opt/test1", "options": "rw,seclabel,relatime,stripe=2048", "size_available": 5005520896, "size_total": 5273976832, "uuid": "72ee3c1e-d5d0-480c-bd29-773c8342d7ac" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.025) 0:02:58.999 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.019) 0:02:59.018 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.022) 0:02:59.041 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.022) 0:02:59.064 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.018) 0:02:59.083 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.019) 0:02:59.102 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.019) 0:02:59.121 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.025) 0:02:59.146 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.019) 0:02:59.165 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.020) 0:02:59.186 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.021) 0:02:59.207 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.019) 0:02:59.226 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 ext2 defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.039) 0:02:59.265 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.022) 0:02:59.288 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.022) 0:02:59.310 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.018) 0:02:59.329 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.017) 0:02:59.347 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.022) 0:02:59.369 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:46:24 +0000 (0:00:00.025) 0:02:59.395 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685587579.4324613, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1685587578.4144647, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1198, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1685587578.4144647, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.208) 0:02:59.603 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.025) 0:02:59.628 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.020) 0:02:59.648 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.024) 0:02:59.672 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.019) 0:02:59.692 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.019) 0:02:59.711 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.022) 0:02:59.734 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:46:25 +0000 (0:00:00.019) 0:02:59.754 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:46:27 +0000 (0:00:02.323) 0:03:02.077 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.020) 0:03:02.098 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.020) 0:03:02.119 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.027) 0:03:02.146 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.021) 0:03:02.167 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.019) 0:03:02.187 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.020) 0:03:02.207 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.021) 0:03:02.229 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.021) 0:03:02.251 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.026) 0:03:02.277 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.025) 0:03:02.302 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.026) 0:03:02.329 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.023) 0:03:02.352 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.059) 0:03:02.412 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.020) 0:03:02.432 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.021) 0:03:02.454 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.020) 0:03:02.474 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.019) 0:03:02.494 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.020) 0:03:02.515 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:46:27 +0000 (0:00:00.023) 0:03:02.538 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.021) 0:03:02.560 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.020) 0:03:02.581 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.204) 0:03:02.785 ********* ok: [sut] => { "bytes": 5368709120, "changed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.204) 0:03:02.990 ********* ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "5368709120" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.027) 0:03:03.017 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.020) 0:03:03.038 ********* ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.202) 0:03:03.240 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.022) 0:03:03.263 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.022) 0:03:03.285 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.023) 0:03:03.309 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.022) 0:03:03.331 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.019) 0:03:03.351 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.019) 0:03:03.370 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.020) 0:03:03.391 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.021) 0:03:03.412 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.020) 0:03:03.433 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.018) 0:03:03.451 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.018) 0:03:03.470 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.018) 0:03:03.488 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.018) 0:03:03.506 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.020) 0:03:03.526 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:46:28 +0000 (0:00:00.023) 0:03:03.549 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.021) 0:03:03.571 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.021) 0:03:03.592 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.020) 0:03:03.613 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.020) 0:03:03.633 ********* ok: [sut] => { "storage_test_actual_size": { "bytes": 5368709120, "changed": false, "failed": false, "lvm": "5g", "parted": "5GiB", "size": "5 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.021) 0:03:03.655 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.024) 0:03:03.679 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.027) 0:03:03.706 ********* ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.034014", "end": "2023-06-01 02:46:29.354528", "rc": 0, "start": "2023-06-01 02:46:29.320514" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.240) 0:03:03.947 ********* ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.026) 0:03:03.973 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.025) 0:03:03.998 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.021) 0:03:04.020 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.020) 0:03:04.040 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.021) 0:03:04.062 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.022) 0:03:04.085 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.021) 0:03:04.107 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.017) 0:03:04.125 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:304 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.019) 0:03:04.144 ********* TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.074) 0:03:04.219 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 01 June 2023 02:46:29 +0000 (0:00:00.026) 0:03:04.245 ********* ok: [sut] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.344) 0:03:04.590 ********* skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_36.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_36.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.049) 0:03:04.639 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.019) 0:03:04.659 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.019) 0:03:04.678 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.040) 0:03:04.718 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:9 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.021) 0:03:04.740 ********* ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "5g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.022) 0:03:04.763 ********* ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.019) 0:03:04.782 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:32 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.019) 0:03:04.801 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.020) 0:03:04.821 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:46 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.019) 0:03:04.841 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:53 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.022) 0:03:04.864 ********* ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:67 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.032) 0:03:04.897 ********* TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Thursday 01 June 2023 02:46:30 +0000 (0:00:00.016) 0:03:04.914 ********* changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "ext2", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:87 Thursday 01 June 2023 02:46:32 +0000 (0:00:02.038) 0:03:06.952 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:99 Thursday 01 June 2023 02:46:32 +0000 (0:00:00.020) 0:03:06.973 ********* TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:105 Thursday 01 June 2023 02:46:32 +0000 (0:00:00.017) 0:03:06.990 ********* ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "ext2" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "ext2", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 01 June 2023 02:46:32 +0000 (0:00:00.022) 0:03:07.012 ********* ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 01 June 2023 02:46:32 +0000 (0:00:00.021) 0:03:07.034 ********* ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:130 Thursday 01 June 2023 02:46:32 +0000 (0:00:00.019) 0:03:07.054 ********* changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'ext2'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "ext2", "mount_info": { "fstype": "ext2", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:142 Thursday 01 June 2023 02:46:32 +0000 (0:00:00.216) 0:03:07.271 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:147 Thursday 01 June 2023 02:46:33 +0000 (0:00:00.609) 0:03:07.880 ********* TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:159 Thursday 01 June 2023 02:46:33 +0000 (0:00:00.018) 0:03:07.899 ********* TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:174 Thursday 01 June 2023 02:46:33 +0000 (0:00:00.018) 0:03:07.918 ********* ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:182 Thursday 01 June 2023 02:46:33 +0000 (0:00:00.593) 0:03:08.511 ********* ok: [sut] => { "changed": false, "stat": { "atime": 1685586918.8447251, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1684244424.757, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1684244183.529, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "3816983141", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:187 Thursday 01 June 2023 02:46:34 +0000 (0:00:00.209) 0:03:08.721 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:209 Thursday 01 June 2023 02:46:34 +0000 (0:00:00.017) 0:03:08.739 ********* ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/tests_resize.yml:317 Thursday 01 June 2023 02:46:34 +0000 (0:00:00.596) 0:03:09.335 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:2 Thursday 01 June 2023 02:46:34 +0000 (0:00:00.051) 0:03:09.387 ********* ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "ext2", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "5g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:7 Thursday 01 June 2023 02:46:34 +0000 (0:00:00.024) 0:03:09.412 ********* skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:15 Thursday 01 June 2023 02:46:34 +0000 (0:00:00.051) 0:03:09.463 ********* ok: [sut] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "f91a7ec7-5021-4d03-b280-c7f5e8053b5f" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:20 Thursday 01 June 2023 02:46:35 +0000 (0:00:00.209) 0:03:09.673 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003388", "end": "2023-06-01 02:46:35.285478", "rc": 0, "start": "2023-06-01 02:46:35.282090" } STDOUT: # # /etc/fstab # Created by anaconda on Tue May 16 13:36:23 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=f91a7ec7-5021-4d03-b280-c7f5e8053b5f / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:25 Thursday 01 June 2023 02:46:35 +0000 (0:00:00.200) 0:03:09.873 ********* ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004653", "end": "2023-06-01 02:46:36.487178", "failed_when_result": false, "rc": 0, "start": "2023-06-01 02:46:35.482525" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:34 Thursday 01 June 2023 02:46:36 +0000 (0:00:01.203) 0:03:11.076 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:5 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.039) 0:03:11.115 ********* ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool.yml:18 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.017) 0:03:11.133 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:2 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.039) 0:03:11.172 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:13 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.025) 0:03:11.197 ********* TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:22 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.015) 0:03:11.213 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:27 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.021) 0:03:11.235 ********* ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:33 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.022) 0:03:11.257 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:42 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.023) 0:03:11.281 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:48 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.022) 0:03:11.304 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:54 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.021) 0:03:11.325 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:59 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.019) 0:03:11.344 ********* TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:73 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.016) 0:03:11.361 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:8 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.039) 0:03:11.400 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:14 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.018) 0:03:11.419 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:21 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.018) 0:03:11.437 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:28 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.019) 0:03:11.456 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:35 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.020) 0:03:11.476 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:45 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.018) 0:03:11.495 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:55 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.020) 0:03:11.515 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-md.yml:66 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.018) 0:03:11.534 ********* ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:76 Thursday 01 June 2023 02:46:36 +0000 (0:00:00.017) 0:03:11.552 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-lvmraid.yml:2 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.036) 0:03:11.588 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:8 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.038) 0:03:11.627 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:16 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.017) 0:03:11.644 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:21 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:11.663 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:29 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:11.681 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:34 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.017) 0:03:11.699 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:40 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.020) 0:03:11.720 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-lvmraid.yml:46 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.020) 0:03:11.740 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:79 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.019) 0:03:11.760 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-thin.yml:2 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.036) 0:03:11.796 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:8 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.040) 0:03:11.837 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:16 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:11.856 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:23 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:11.874 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-thin.yml:27 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:11.892 ********* ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:82 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:11.910 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:5 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.041) 0:03:11.952 ********* ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:13 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.020) 0:03:11.972 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:20 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.015) 0:03:11.988 ********* TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-encryption.yml:27 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.015) 0:03:12.004 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:85 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.017) 0:03:12.021 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-members-vdo.yml:2 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.042) 0:03:12.064 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:9 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.038) 0:03:12.102 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:16 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.020) 0:03:12.123 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:22 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.049) 0:03:12.173 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:28 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.020) 0:03:12.193 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:35 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.019) 0:03:12.212 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:41 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:12.230 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-pool-member-vdo.yml:47 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.019) 0:03:12.250 ********* ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-members.yml:88 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.018) 0:03:12.268 ********* ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-pool-volumes.yml:3 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.019) 0:03:12.287 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:2 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.035) 0:03:12.322 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:21 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.022) 0:03:12.345 ********* included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:7 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.093) 0:03:12.439 ********* ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:16 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.022) 0:03:12.461 ********* ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:33 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.025) 0:03:12.486 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:46 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.019) 0:03:12.505 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:58 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.017) 0:03:12.523 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:66 Thursday 01 June 2023 02:46:37 +0000 (0:00:00.021) 0:03:12.545 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:78 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.020) 0:03:12.565 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:90 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.584 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:105 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.603 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:117 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.622 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:123 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.019) 0:03:12.641 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:129 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.659 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-mount.yml:141 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.019) 0:03:12.679 ********* ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:2 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.017) 0:03:12.697 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:40 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.036) 0:03:12.733 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:48 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.752 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:58 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.022) 0:03:12.775 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fstab.yml:71 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.020) 0:03:12.795 ********* ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:3 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.017) 0:03:12.813 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-fs.yml:10 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.831 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:3 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.018) 0:03:12.849 ********* ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:9 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.200) 0:03:13.050 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:16 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.019) 0:03:13.069 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:24 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.023) 0:03:13.092 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:30 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.015) 0:03:13.108 ********* ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:34 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.021) 0:03:13.129 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-device.yml:39 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.019) 0:03:13.149 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:3 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.017) 0:03:13.166 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Thursday 01 June 2023 02:46:38 +0000 (0:00:00.019) 0:03:13.186 ********* ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:15 Thursday 01 June 2023 02:46:40 +0000 (0:00:02.334) 0:03:15.521 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:21 Thursday 01 June 2023 02:46:40 +0000 (0:00:00.019) 0:03:15.540 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:30 Thursday 01 June 2023 02:46:40 +0000 (0:00:00.018) 0:03:15.558 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:43 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.015) 0:03:15.574 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:49 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.592 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:54 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.611 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:67 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.016) 0:03:15.628 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:79 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.014) 0:03:15.642 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:92 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.015) 0:03:15.658 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:104 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.025) 0:03:15.683 ********* ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:112 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.021) 0:03:15.705 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:120 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.723 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:129 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.019) 0:03:15.743 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:138 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.762 ********* ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:8 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.780 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:14 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:15.801 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:21 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.820 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:28 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.838 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:35 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.856 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:44 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.875 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-md.yml:53 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:15.893 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:3 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.021) 0:03:15.915 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:11 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.056) 0:03:15.971 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:20 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:15.992 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:28 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.019) 0:03:16.011 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:32 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:16.031 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:46 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.050 ********* skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:50 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.068 ********* skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:54 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:16.088 ********* skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:58 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.107 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:68 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.125 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:72 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.144 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:77 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.017) 0:03:16.161 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:83 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.017) 0:03:16.179 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:88 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.019) 0:03:16.199 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:96 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.017) 0:03:16.216 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:104 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.017) 0:03:16.234 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:109 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.253 ********* skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:113 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.271 ********* skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:117 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.290 ********* skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:121 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:16.311 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:129 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.019) 0:03:16.330 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:138 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.349 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:142 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.367 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:150 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.386 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:156 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.405 ********* ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:160 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:16.426 ********* ok: [sut] => { "storage_test_expected_size": "5368709120" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-size.yml:164 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.444 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:5 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.017) 0:03:16.462 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:13 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.017) 0:03:16.480 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:18 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.498 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:27 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.018) 0:03:16.516 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:35 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.020) 0:03:16.536 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:41 Thursday 01 June 2023 02:46:41 +0000 (0:00:00.019) 0:03:16.556 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-cache.yml:47 Thursday 01 June 2023 02:46:42 +0000 (0:00:00.018) 0:03:16.574 ********* skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume.yml:27 Thursday 01 June 2023 02:46:42 +0000 (0:00:00.018) 0:03:16.592 ********* ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:44 Thursday 01 June 2023 02:46:42 +0000 (0:00:00.017) 0:03:16.609 ********* TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/verify-role-results.yml:54 Thursday 01 June 2023 02:46:42 +0000 (0:00:00.015) 0:03:16.625 ********* ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* sut : ok=1270 changed=28 unreachable=0 failed=3 skipped=1281 rescued=3 ignored=0 Thursday 01 June 2023 02:46:42 +0000 (0:00:00.008) 0:03:16.634 ********* =============================================================================== linux-system-roles.storage : Make sure blivet is available -------------- 7.58s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 3.63s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Ensure cryptsetup is present -------------------------------------------- 3.04s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 2.79s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 2.78s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 2.42s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 2.42s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:73 Ensure cryptsetup is present -------------------------------------------- 2.37s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.37s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 linux-system-roles.storage : Make sure required packages are installed --- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:39 Ensure cryptsetup is present -------------------------------------------- 2.34s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.33s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.32s /WORKDIR/git-main_lvm-raid-stripe-sizeq7sss_ht/tests/test-verify-volume-encryption.yml:10 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- /opt/ansible-2.9/lib/python3.6/site-packages/ansible/parsing/vault/__init__.py:44: CryptographyDeprecationWarning: Python 3.6 is no longer supported by the Python core team. Therefore, support for it is deprecated in cryptography. The next release of cryptography will remove support for Python 3.6. from cryptography.exceptions import InvalidSignature [DEPRECATION WARNING]: Distribution fedora 36 on host sut should use /usr/bin/python3, but is using /usr/bin/python for backward compatibility with prior Ansible releases. A future Ansible release will default to using the discovered platform python for this host. See https://docs.ansible.com/ansible/ 2.9/reference_appendices/interpreter_discovery.html for more information. This feature will be removed in version 2.12. Deprecation warnings can be disabled by setting deprecation_warnings=False in ansible.cfg. ---^---^---^---^---^---