# STDOUT: ---v---v---v---v---v--- ansible-playbook 2.9.27 config file = /etc/ansible/ansible.cfg configured module search path = ['/home/jenkins/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /opt/ansible-2.9/lib/python3.9/site-packages/ansible executable location = /opt/ansible-2.9/bin/ansible-playbook python version = 3.9.18 (main, Sep 7 2023, 00:00:00) [GCC 11.4.1 20230605 (Red Hat 11.4.1-2)] Using /etc/ansible/ansible.cfg as config file statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml statically imported: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml Skipping callback 'actionable', as we already have a stdout callback. Skipping callback 'counter_enabled', as we already have a stdout callback. Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'dense', as we already have a stdout callback. Skipping callback 'full_skip', as we already have a stdout callback. Skipping callback 'json', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'null', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. Skipping callback 'selective', as we already have a stdout callback. Skipping callback 'skippy', as we already have a stdout callback. Skipping callback 'stderr', as we already have a stdout callback. Skipping callback 'unixy', as we already have a stdout callback. Skipping callback 'yaml', as we already have a stdout callback. PLAYBOOK: tests_luks.yml ******************************************************* 1 plays in /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml PLAY [Test LUKS] *************************************************************** TASK [Gathering Facts] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:2 Thursday 23 November 2023 00:22:27 +0000 (0:00:00.020) 0:00:00.020 ***** ok: [sut] META: ran handlers TASK [Enable FIPS mode] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:21 Thursday 23 November 2023 00:22:30 +0000 (0:00:02.784) 0:00:02.804 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reboot] ****************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:25 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.025) 0:00:02.830 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure dracut-fips] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:35 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.025) 0:00:02.856 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Configure boot for FIPS] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:42 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.026) 0:00:02.882 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reboot] ****************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:51 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.022) 0:00:02.905 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Run the role] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:55 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.022) 0:00:02.927 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.026) 0:00:02.954 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.028) 0:00:02.983 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.022) 0:00:03.006 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.042) 0:00:03.048 ***** ok: [sut] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.299) 0:00:03.348 ***** ok: [sut] => { "ansible_facts": { "__storage_is_ostree": false }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.024) 0:00:03.373 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.011) 0:00:03.384 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.009) 0:00:03.394 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:22:30 +0000 (0:00:00.037) 0:00:03.432 ***** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: lsof-4.96.3-3.fc38.x86_64", "Installed: python3-blivet-1:3.7.1-1.fc38.noarch", "Installed: python3-blockdev-2.28-5.fc38.x86_64", "Installed: libblockdev-btrfs-2.28-5.fc38.x86_64", "Installed: python3-pyparted-1:3.12.0-8.fc38.x86_64", "Installed: btrfs-progs-6.5.1-1.fc38.x86_64", "Installed: blivet-data-1:3.7.1-1.fc38.noarch", "Installed: device-mapper-multipath-0.9.4-2.fc38.x86_64", "Installed: libblockdev-mpath-2.28-5.fc38.x86_64", "Installed: device-mapper-multipath-libs-0.9.4-2.fc38.x86_64", "Installed: libblockdev-dm-2.28-5.fc38.x86_64", "Installed: lzo-2.10-8.fc38.x86_64", "Installed: python3-bytesize-2.10-1.fc38.x86_64", "Installed: libblockdev-lvm-2.28-5.fc38.x86_64" ] } TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:22:36 +0000 (0:00:06.203) 0:00:09.635 ***** ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:22:36 +0000 (0:00:00.025) 0:00:09.661 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:22:36 +0000 (0:00:00.024) 0:00:09.685 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:22:37 +0000 (0:00:00.623) 0:00:10.309 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:22:37 +0000 (0:00:00.042) 0:00:10.351 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:22:37 +0000 (0:00:00.020) 0:00:10.372 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:22:37 +0000 (0:00:00.011) 0:00:10.383 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:22:37 +0000 (0:00:00.020) 0:00:10.403 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:22:40 +0000 (0:00:02.450) 0:00:12.854 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "inactive", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:22:42 +0000 (0:00:02.499) 0:00:15.354 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:22:42 +0000 (0:00:00.029) 0:00:15.384 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:22:42 +0000 (0:00:00.014) 0:00:15.398 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.381) 0:00:15.780 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.015) 0:00:15.795 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.012) 0:00:15.808 ***** ok: [sut] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.020) 0:00:15.828 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.016) 0:00:15.844 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.014) 0:00:15.859 ***** TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.013) 0:00:15.873 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.013) 0:00:15.886 ***** TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.013) 0:00:15.900 ***** TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.014) 0:00:15.914 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.013) 0:00:15.928 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700075992.528, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700076436.853, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1700075992.528, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "550502605", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.215) 0:00:16.143 ***** TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:22:43 +0000 (0:00:00.017) 0:00:16.160 ***** ok: [sut] TASK [Get unused disks] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:59 Thursday 23 November 2023 00:22:44 +0000 (0:00:00.803) 0:00:16.964 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/get_unused_disk.yml for sut TASK [Ensure test packages] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/get_unused_disk.yml:2 Thursday 23 November 2023 00:22:44 +0000 (0:00:00.024) 0:00:16.989 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Find unused disks in the system] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/get_unused_disk.yml:16 Thursday 23 November 2023 00:22:46 +0000 (0:00:02.485) 0:00:19.474 ***** ok: [sut] => { "changed": false, "disks": [ "sda" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/get_unused_disk.yml:24 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.304) 0:00:19.778 ***** ok: [sut] => { "ansible_facts": { "unused_disks": [ "sda" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/get_unused_disk.yml:29 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.017) 0:00:19.796 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/get_unused_disk.yml:34 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.013) 0:00:19.810 ***** ok: [sut] => { "unused_disks": [ "sda" ] } TASK [Test for correct handling of new encrypted volume w/ no key] ************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:68 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.014) 0:00:19.824 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.025) 0:00:19.849 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.017) 0:00:19.867 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.040) 0:00:19.908 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.021) 0:00:19.929 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.016) 0:00:19.945 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.036) 0:00:19.982 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.013) 0:00:19.995 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.013) 0:00:20.009 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.012) 0:00:20.022 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.012) 0:00:20.034 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:22:47 +0000 (0:00:00.032) 0:00:20.067 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:22:49 +0000 (0:00:02.546) 0:00:22.613 ***** ok: [sut] => { "storage_pools": [] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:22:49 +0000 (0:00:00.019) 0:00:22.633 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "encryption": true, "mount_point": "/opt/test1", "name": "foo", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:22:49 +0000 (0:00:00.018) 0:00:22.652 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:22:51 +0000 (0:00:01.904) 0:00:24.556 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:22:51 +0000 (0:00:00.028) 0:00:24.585 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:22:51 +0000 (0:00:00.027) 0:00:24.612 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:22:51 +0000 (0:00:00.019) 0:00:24.632 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:22:51 +0000 (0:00:00.028) 0:00:24.661 ***** changed: [sut] => { "changed": true, "rc": 0, "results": [ "Installed: cryptsetup-2.6.1-1.fc38.x86_64" ] } TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:22:55 +0000 (0:00:03.514) 0:00:28.175 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "run-r84fb8ce5825044709a6ad664f237e420.service": { "name": "run-r84fb8ce5825044709a6ad664f237e420.service", "source": "systemd", "state": "running", "status": "transient" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:22:57 +0000 (0:00:02.445) 0:00:30.621 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:22:57 +0000 (0:00:00.024) 0:00:30.645 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:22:57 +0000 (0:00:00.038) 0:00:30.683 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: encrypted volume 'foo' missing key/password TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:22:59 +0000 (0:00:01.628) 0:00:32.311 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "encrypted volume 'foo' missing key/password", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [], 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'foo', 'raid_level': None, 'size': 10737418240, 'state': 'present', 'type': 'disk', 'disks': ['sda'], 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_stripe_size': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.020) 0:00:32.331 ***** TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.014) 0:00:32.346 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.018) 0:00:32.364 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.022) 0:00:32.386 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Create an encrypted disk volume w/ default fs] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:83 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.015) 0:00:32.402 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.038) 0:00:32.440 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.022) 0:00:32.462 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.019) 0:00:32.481 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.040) 0:00:32.521 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.014) 0:00:32.536 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.014) 0:00:32.550 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.013) 0:00:32.563 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.012) 0:00:32.576 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:22:59 +0000 (0:00:00.031) 0:00:32.607 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:23:02 +0000 (0:00:02.549) 0:00:35.157 ***** ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:23:02 +0000 (0:00:00.015) 0:00:35.172 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "foo", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:23:02 +0000 (0:00:00.017) 0:00:35.190 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:23:03 +0000 (0:00:01.559) 0:00:36.749 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:23:04 +0000 (0:00:00.026) 0:00:36.776 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:23:04 +0000 (0:00:00.022) 0:00:36.799 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:23:04 +0000 (0:00:00.014) 0:00:36.813 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:23:04 +0000 (0:00:00.027) 0:00:36.841 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:23:06 +0000 (0:00:02.520) 0:00:39.361 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:23:09 +0000 (0:00:02.480) 0:00:41.841 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:23:09 +0000 (0:00:00.069) 0:00:41.911 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:23:09 +0000 (0:00:00.015) 0:00:41.926 ***** changed: [sut] => { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "password": "-", "state": "present" } ], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "mounted" } ], "packages": [ "xfsprogs", "cryptsetup", "e2fsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:23:20 +0000 (0:00:10.858) 0:00:52.784 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:23:20 +0000 (0:00:00.016) 0:00:52.801 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:23:20 +0000 (0:00:00.015) 0:00:52.816 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "mounted" } ], "packages": [ "xfsprogs", "cryptsetup", "e2fsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:23:20 +0000 (0:00:00.027) 0:00:52.844 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:23:20 +0000 (0:00:00.018) 0:00:52.862 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:23:20 +0000 (0:00:00.020) 0:00:52.883 ***** TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:23:20 +0000 (0:00:00.016) 0:00:52.899 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:23:21 +0000 (0:00:01.180) 0:00:54.079 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:23:22 +0000 (0:00:01.142) 0:00:55.222 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:23:22 +0000 (0:00:00.023) 0:00:55.245 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:23:23 +0000 (0:00:01.386) 0:00:56.631 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700698963.3445008, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700076436.853, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131081, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1700075992.528, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "550502605", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:23:24 +0000 (0:00:00.250) 0:00:56.881 ***** changed: [sut] => (item={'backing_device': '/dev/sda', 'name': 'luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47', 'password': '-', 'state': 'present'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda", "name": "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "password": "-", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:23:24 +0000 (0:00:00.442) 0:00:57.324 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:95 Thursday 23 November 2023 00:23:25 +0000 (0:00:01.084) 0:00:58.408 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:23:25 +0000 (0:00:00.029) 0:00:58.438 ***** skipping: [sut] => {} TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:23:25 +0000 (0:00:00.014) 0:00:58.453 ***** ok: [sut] => { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:23:25 +0000 (0:00:00.017) 0:00:58.470 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "size": "10G", "type": "crypt", "uuid": "b8f79632-3c7e-4a8b-a92e-9f36a97b087c" }, "/dev/sda": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "2e45363f-c6b9-4108-80da-9b1e68fd9f47" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.312) 0:00:58.783 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003449", "end": "2023-11-23 00:23:26.284518", "rc": 0, "start": "2023-11-23 00:23:26.281069" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.312) 0:00:59.096 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003340", "end": "2023-11-23 00:23:26.519533", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:23:26.516193" } STDOUT: luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 /dev/sda - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.217) 0:00:59.314 ***** TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.020) 0:00:59.334 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.035) 0:00:59.370 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.020) 0:00:59.390 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.086) 0:00:59.476 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.022) 0:00:59.498 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2574562, "block_size": 4096, "block_total": 2600960, "block_used": 26398, "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fstype": "xfs", "inode_available": 5234685, "inode_total": 5234688, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10545405952, "size_total": 10653532160, "uuid": "b8f79632-3c7e-4a8b-a92e-9f36a97b087c" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2574562, "block_size": 4096, "block_total": 2600960, "block_used": 26398, "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fstype": "xfs", "inode_available": 5234685, "inode_total": 5234688, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10545405952, "size_total": 10653532160, "uuid": "b8f79632-3c7e-4a8b-a92e-9f36a97b087c" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.025) 0:00:59.524 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.016) 0:00:59.540 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.020) 0:00:59.561 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.023) 0:00:59.584 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.016) 0:00:59.600 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.015) 0:00:59.616 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.016) 0:00:59.632 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.018) 0:00:59.651 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.014) 0:00:59.666 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.014) 0:00:59.681 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.014) 0:00:59.695 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.014) 0:00:59.709 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:23:26 +0000 (0:00:00.035) 0:00:59.745 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.022) 0:00:59.767 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.017) 0:00:59.785 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.014) 0:00:59.800 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.014) 0:00:59.814 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.020) 0:00:59.834 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.021) 0:00:59.855 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699004.5139499, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700698999.6011353, "dev": 5, "device_type": 2048, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 574, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700698999.6011353, "nlink": 1, "path": "/dev/sda", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.215) 0:01:00.071 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.021) 0:01:00.093 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.015) 0:01:00.109 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.017) 0:01:00.127 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "disk" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.015) 0:01:00.142 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.013) 0:01:00.156 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.015) 0:01:00.172 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699004.5159497, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700698999.8911245, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 776, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700698999.8911245, "nlink": 1, "path": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:23:27 +0000 (0:00:00.219) 0:01:00.391 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:23:30 +0000 (0:00:02.748) 0:01:03.140 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/sda" ], "delta": "0:00:00.008529", "end": "2023-11-23 00:23:30.572262", "rc": 0, "start": "2023-11-23 00:23:30.563733" } STDOUT: LUKS header information Version: 2 Epoch: 3 Metadata area: 16384 [bytes] Keyslots area: 16744448 [bytes] UUID: 2e45363f-c6b9-4108-80da-9b1e68fd9f47 Label: (no label) Subsystem: (no subsystem) Flags: (no flags) Data segments: 0: crypt offset: 16777216 [bytes] length: (whole device) cipher: aes-xts-plain64 sector: 512 [bytes] Keyslots: 0: luks2 Key: 512 bits Priority: normal Cipher: aes-xts-plain64 Cipher key: 512 bits PBKDF: argon2id Time cost: 4 Memory: 706690 Threads: 2 Salt: 73 3d 6d cd a0 09 2f f1 bf 88 db d9 26 6d 4a 87 7e 70 23 7a 61 87 4a 2f e3 1c c9 a4 4f 9a 05 6f AF stripes: 4000 AF hash: sha256 Area offset:32768 [bytes] Area length:258048 [bytes] Digest ID: 0 Tokens: Digests: 0: pbkdf2 Hash: sha256 Iterations: 92827 Salt: 50 25 b9 70 d8 83 33 19 68 27 4c f8 28 a1 55 9b d6 12 9b 5b 5e e8 60 a2 3c 59 77 da e1 74 de 11 Digest: 57 91 b1 77 7b 15 28 c0 98 f8 d1 77 05 26 1a d0 6f 7d 4c 28 42 c5 66 9f 0f cc 10 79 c7 bb fd c4 TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.223) 0:01:03.363 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.020) 0:01:03.383 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.021) 0:01:03.405 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.018) 0:01:03.423 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.018) 0:01:03.441 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.014) 0:01:03.456 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.014) 0:01:03.471 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.014) 0:01:03.486 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 /dev/sda -" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.020) 0:01:03.507 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.017) 0:01:03.524 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.018) 0:01:03.542 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.018) 0:01:03.560 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.018) 0:01:03.579 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.012) 0:01:03.591 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.015) 0:01:03.607 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.621 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.634 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.648 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.662 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.676 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.014) 0:01:03.691 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.704 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.718 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.013) 0:01:03.731 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.015) 0:01:03.747 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:23:30 +0000 (0:00:00.014) 0:01:03.761 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.016) 0:01:03.777 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.792 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.807 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.822 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.836 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.851 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.016) 0:01:03.867 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:03.881 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.895 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.910 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.925 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:03.939 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:03.954 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:03.968 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:03.983 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:03.996 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:04.010 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:04.024 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:04.039 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:04.053 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:04.067 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:04.081 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:04.096 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:04.111 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.016) 0:01:04.128 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.032) 0:01:04.160 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:04.176 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.013) 0:01:04.190 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:04.204 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.016) 0:01:04.221 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:04.237 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.021) 0:01:04.258 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.016) 0:01:04.275 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml:12 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.015) 0:01:04.290 ***** changed: [sut] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Test for correct handling of safe_mode] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:101 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.285) 0:01:04.576 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.030) 0:01:04.607 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.019) 0:01:04.626 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.021) 0:01:04.648 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.023) 0:01:04.672 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.019) 0:01:04.692 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.038) 0:01:04.730 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:04.744 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:23:31 +0000 (0:00:00.014) 0:01:04.758 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:23:32 +0000 (0:00:00.012) 0:01:04.771 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:23:32 +0000 (0:00:00.013) 0:01:04.784 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:23:32 +0000 (0:00:00.031) 0:01:04.816 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:23:34 +0000 (0:00:02.507) 0:01:07.323 ***** ok: [sut] => { "storage_pools": [] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:23:34 +0000 (0:00:00.019) 0:01:07.343 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "encryption": false, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "foo", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:23:34 +0000 (0:00:00.020) 0:01:07.364 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:23:36 +0000 (0:00:01.957) 0:01:09.321 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:23:36 +0000 (0:00:00.028) 0:01:09.349 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:23:36 +0000 (0:00:00.022) 0:01:09.371 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:23:36 +0000 (0:00:00.014) 0:01:09.386 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:23:36 +0000 (0:00:00.022) 0:01:09.408 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:23:39 +0000 (0:00:02.505) 0:01:11.913 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:23:41 +0000 (0:00:02.427) 0:01:14.341 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:23:41 +0000 (0:00:00.024) 0:01:14.365 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:23:41 +0000 (0:00:00.012) 0:01:14.378 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting on device 'luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47' in safe mode due to encryption removal TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:23:43 +0000 (0:00:01.763) 0:01:16.141 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting on device 'luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47' in safe mode due to encryption removal", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [], 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'foo', 'raid_level': None, 'size': 10720641024, 'state': 'present', 'type': 'disk', 'disks': ['sda'], 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_stripe_size': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.024) 0:01:16.165 ***** TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.014) 0:01:16.180 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.017) 0:01:16.197 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.025) 0:01:16.222 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:11 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.015) 0:01:16.238 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699011.7866719, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699011.7866719, "dev": 64768, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1700699011.7866719, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "2069102140", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Assert file presence] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:16 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.221) 0:01:16.459 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Remove the encryption layer] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:121 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.017) 0:01:16.477 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.040) 0:01:16.518 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.021) 0:01:16.539 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.017) 0:01:16.556 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.037) 0:01:16.594 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.014) 0:01:16.608 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.013) 0:01:16.622 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.013) 0:01:16.636 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.012) 0:01:16.649 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:23:43 +0000 (0:00:00.030) 0:01:16.679 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:23:46 +0000 (0:00:02.577) 0:01:19.257 ***** ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:23:46 +0000 (0:00:00.017) 0:01:19.274 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "encryption": false, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "foo", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:23:46 +0000 (0:00:00.020) 0:01:19.294 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:23:48 +0000 (0:00:01.720) 0:01:21.015 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:23:48 +0000 (0:00:00.028) 0:01:21.044 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:23:48 +0000 (0:00:00.025) 0:01:21.070 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:23:48 +0000 (0:00:00.015) 0:01:21.085 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:23:48 +0000 (0:00:00.022) 0:01:21.108 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:23:50 +0000 (0:00:02.522) 0:01:23.631 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:23:53 +0000 (0:00:02.394) 0:01:26.025 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:23:53 +0000 (0:00:00.024) 0:01:26.050 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:23:53 +0000 (0:00:00.015) 0:01:26.065 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create format", "device": "/dev/sda", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/sda", "_kernel_device": "/dev/sda", "_mount_id": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10720641024, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:23:55 +0000 (0:00:02.269) 0:01:28.334 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:23:55 +0000 (0:00:00.014) 0:01:28.349 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:23:55 +0000 (0:00:00.012) 0:01:28.362 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create format", "device": "/dev/sda", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/sda", "_kernel_device": "/dev/sda", "_mount_id": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10720641024, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:23:55 +0000 (0:00:00.017) 0:01:28.379 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:23:55 +0000 (0:00:00.017) 0:01:28.397 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/sda", "_kernel_device": "/dev/sda", "_mount_id": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10720641024, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:23:55 +0000 (0:00:00.016) 0:01:28.413 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:23:55 +0000 (0:00:00.227) 0:01:28.641 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:23:56 +0000 (0:00:01.012) 0:01:29.653 ***** changed: [sut] => (item={'src': 'UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:23:57 +0000 (0:00:00.256) 0:01:29.910 ***** skipping: [sut] => (item={'src': 'UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:23:57 +0000 (0:00:00.022) 0:01:29.932 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:23:58 +0000 (0:00:00.996) 0:01:30.929 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699004.51195, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "9bcba2755e296bc70b1ae13c4fc4022db829d809", "ctime": 1700699004.50895, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699004.50795, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 53, "uid": 0, "version": "3285124333", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:23:58 +0000 (0:00:00.223) 0:01:31.152 ***** changed: [sut] => (item={'backing_device': '/dev/sda', 'name': 'luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47', 'password': '-', 'state': 'absent'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda", "name": "luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:23:58 +0000 (0:00:00.238) 0:01:31.391 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:134 Thursday 23 November 2023 00:23:59 +0000 (0:00:00.815) 0:01:32.207 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:23:59 +0000 (0:00:00.031) 0:01:32.238 ***** skipping: [sut] => {} TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:23:59 +0000 (0:00:00.013) 0:01:32.252 ***** ok: [sut] => { "_storage_volumes_list": [ { "_device": "/dev/sda", "_kernel_device": "/dev/sda", "_mount_id": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10720641024, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:23:59 +0000 (0:00:00.017) 0:01:32.270 ***** ok: [sut] => { "changed": false, "info": { "/dev/sda": { "fstype": "xfs", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "33d02b0c-d075-4952-9a0a-d69e26c2076a" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:23:59 +0000 (0:00:00.215) 0:01:32.485 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003498", "end": "2023-11-23 00:23:59.907008", "rc": 0, "start": "2023-11-23 00:23:59.903510" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:23:59 +0000 (0:00:00.213) 0:01:32.698 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003505", "end": "2023-11-23 00:24:00.124727", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:24:00.121222" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.216) 0:01:32.915 ***** TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.013) 0:01:32.929 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.030) 0:01:32.959 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.019) 0:01:32.979 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.071) 0:01:33.050 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/sda" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.019) 0:01:33.070 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2578630, "block_size": 4096, "block_total": 2605056, "block_used": 26426, "device": "/dev/sda", "fstype": "xfs", "inode_available": 5242877, "inode_total": 5242880, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10562068480, "size_total": 10670309376, "uuid": "33d02b0c-d075-4952-9a0a-d69e26c2076a" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2578630, "block_size": 4096, "block_total": 2605056, "block_used": 26426, "device": "/dev/sda", "fstype": "xfs", "inode_available": 5242877, "inode_total": 5242880, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10562068480, "size_total": 10670309376, "uuid": "33d02b0c-d075-4952-9a0a-d69e26c2076a" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.024) 0:01:33.094 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.015) 0:01:33.110 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.020) 0:01:33.130 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.019) 0:01:33.150 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.016) 0:01:33.167 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.016) 0:01:33.183 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.015) 0:01:33.199 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.020) 0:01:33.219 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.014) 0:01:33.234 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.014) 0:01:33.249 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.015) 0:01:33.264 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.015) 0:01:33.280 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.065) 0:01:33.345 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.020) 0:01:33.365 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.020) 0:01:33.385 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.018) 0:01:33.404 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.017) 0:01:33.422 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.022) 0:01:33.444 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.020) 0:01:33.464 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699038.5876472, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699035.476766, "dev": 5, "device_type": 2048, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 574, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700699035.476766, "nlink": 1, "path": "/dev/sda", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.220) 0:01:33.685 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.020) 0:01:33.706 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.015) 0:01:33.721 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.018) 0:01:33.740 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "disk" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:24:00 +0000 (0:00:00.015) 0:01:33.755 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:24:01 +0000 (0:00:00.014) 0:01:33.769 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:24:01 +0000 (0:00:00.016) 0:01:33.786 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:24:01 +0000 (0:00:00.015) 0:01:33.801 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:24:03 +0000 (0:00:02.528) 0:01:36.330 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.016) 0:01:36.346 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.014) 0:01:36.360 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.021) 0:01:36.382 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.014) 0:01:36.397 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.413 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.016) 0:01:36.429 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.014) 0:01:36.443 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.459 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.019) 0:01:36.479 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.017) 0:01:36.496 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.510 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.526 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.539 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.012) 0:01:36.552 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.565 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.579 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.593 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.014) 0:01:36.608 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.621 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.635 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.649 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.662 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.013) 0:01:36.676 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.692 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.708 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.723 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.016) 0:01:36.739 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:24:03 +0000 (0:00:00.015) 0:01:36.754 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.015) 0:01:36.769 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.021) 0:01:36.791 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.018) 0:01:36.810 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.015) 0:01:36.826 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.016) 0:01:36.842 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.856 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.869 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.015) 0:01:36.884 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.034) 0:01:36.919 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.933 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.946 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.960 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.973 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:36.986 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.014) 0:01:37.001 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.014 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.028 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.041 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.055 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.068 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.015) 0:01:37.084 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.014) 0:01:37.099 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.015) 0:01:37.114 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.128 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.142 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.156 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.014) 0:01:37.171 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.184 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.014) 0:01:37.198 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.014) 0:01:37.213 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.226 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml:12 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.240 ***** changed: [sut] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Test for correct handling of safe_mode] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:140 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.224) 0:01:37.465 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.034) 0:01:37.499 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.018) 0:01:37.518 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.021) 0:01:37.540 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.022) 0:01:37.562 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.016) 0:01:37.578 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.037) 0:01:37.616 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.630 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.015) 0:01:37.645 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.659 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.013) 0:01:37.672 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:24:04 +0000 (0:00:00.031) 0:01:37.703 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:24:07 +0000 (0:00:02.528) 0:01:40.232 ***** ok: [sut] => { "storage_pools": [] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:24:07 +0000 (0:00:00.018) 0:01:40.251 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "foo", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:24:07 +0000 (0:00:00.039) 0:01:40.290 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:24:09 +0000 (0:00:01.624) 0:01:41.915 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:24:09 +0000 (0:00:00.029) 0:01:41.944 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:24:09 +0000 (0:00:00.024) 0:01:41.968 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:24:09 +0000 (0:00:00.014) 0:01:41.983 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:24:09 +0000 (0:00:00.022) 0:01:42.005 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:24:11 +0000 (0:00:02.503) 0:01:44.509 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service": { "name": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:24:14 +0000 (0:00:02.485) 0:01:46.994 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:24:14 +0000 (0:00:00.025) 0:01:47.019 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d2e45363f\x2dc6b9\x2d4108\x2d80da\x2d9b1e68fd9f47.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "name": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-udevd-kernel.socket dev-sda.device systemd-journald.socket \"system-systemd\\\\x2dcryptsetup.slice\" cryptsetup-pre.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "\"blockdev@dev-mapper-luks\\\\x2d2e45363f\\\\x2dc6b9\\\\x2d4108\\\\x2d80da\\\\x2d9b1e68fd9f47.target\" umount.target cryptsetup.target", "BindsTo": "dev-sda.device", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 /dev/sda - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 /dev/sda - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-2e45363f-c6b9-4108-80da-9b1e68fd9f47 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d2e45363f\\\\x2dc6b9\\\\x2d4108\\\\x2d80da\\\\x2d9b1e68fd9f47.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:23:58 UTC", "StateChangeTimestampMonotonic": "973021726", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d2e45363f\\\\x2dc6b9\\\\x2d4108\\\\x2d80da\\\\x2d9b1e68fd9f47.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:24:15 +0000 (0:00:01.035) 0:01:48.055 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting on device 'sda' in safe mode due to adding encryption TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:24:16 +0000 (0:00:01.619) 0:01:49.674 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting on device 'sda' in safe mode due to adding encryption", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [], 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'foo', 'raid_level': None, 'size': 10737418240, 'state': 'present', 'type': 'disk', 'disks': ['sda'], 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_stripe_size': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:24:16 +0000 (0:00:00.018) 0:01:49.693 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d2e45363f\x2dc6b9\x2d4108\x2d80da\x2d9b1e68fd9f47.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "name": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d2e45363f\\x2dc6b9\\x2d4108\\x2d80da\\x2d9b1e68fd9f47.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d2e45363f\\\\x2dc6b9\\\\x2d4108\\\\x2d80da\\\\x2d9b1e68fd9f47.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:24:17 +0000 (0:00:01.029) 0:01:50.722 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:24:17 +0000 (0:00:00.018) 0:01:50.740 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.023) 0:01:50.764 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:11 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.014) 0:01:50.779 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699044.6724145, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699044.6724145, "dev": 2048, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1700699044.6724145, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "2028416660", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Assert file presence] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:16 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.220) 0:01:50.999 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Add encryption to the volume] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:160 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.018) 0:01:51.017 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.051) 0:01:51.069 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.047) 0:01:51.116 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.017) 0:01:51.134 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.037) 0:01:51.172 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.014) 0:01:51.186 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.013) 0:01:51.199 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.014) 0:01:51.214 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.012) 0:01:51.227 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:24:18 +0000 (0:00:00.030) 0:01:51.257 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:24:20 +0000 (0:00:02.492) 0:01:53.750 ***** ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:24:21 +0000 (0:00:00.016) 0:01:53.767 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "foo", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:24:21 +0000 (0:00:00.018) 0:01:53.785 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:24:22 +0000 (0:00:01.596) 0:01:55.381 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:24:22 +0000 (0:00:00.028) 0:01:55.410 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:24:22 +0000 (0:00:00.025) 0:01:55.435 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:24:22 +0000 (0:00:00.015) 0:01:55.451 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:24:22 +0000 (0:00:00.023) 0:01:55.474 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:24:25 +0000 (0:00:02.516) 0:01:57.991 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:24:27 +0000 (0:00:02.417) 0:02:00.408 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:24:27 +0000 (0:00:00.024) 0:02:00.432 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:24:27 +0000 (0:00:00.012) 0:02:00.445 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/sda", "fs_type": "xfs" }, { "action": "create format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "password": "-", "state": "present" } ], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "mounted" } ], "packages": [ "cryptsetup", "e2fsprogs", "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:24:38 +0000 (0:00:11.222) 0:02:11.667 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:24:38 +0000 (0:00:00.017) 0:02:11.685 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:24:38 +0000 (0:00:00.016) 0:02:11.701 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/sda", "fs_type": "xfs" }, { "action": "create format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "mounted" } ], "packages": [ "cryptsetup", "e2fsprogs", "xfsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:24:38 +0000 (0:00:00.020) 0:02:11.722 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:24:38 +0000 (0:00:00.017) 0:02:11.740 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:24:38 +0000 (0:00:00.018) 0:02:11.759 ***** changed: [sut] => (item={'src': 'UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "UUID=33d02b0c-d075-4952-9a0a-d69e26c2076a" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:24:39 +0000 (0:00:00.232) 0:02:11.991 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:24:40 +0000 (0:00:01.005) 0:02:12.996 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:24:40 +0000 (0:00:00.248) 0:02:13.245 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:24:40 +0000 (0:00:00.020) 0:02:13.265 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:24:41 +0000 (0:00:01.019) 0:02:14.285 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699040.1235883, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699038.5836473, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263320, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1700699038.5816474, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1869261595", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:24:41 +0000 (0:00:00.222) 0:02:14.508 ***** changed: [sut] => (item={'backing_device': '/dev/sda', 'name': 'luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf', 'password': '-', 'state': 'present'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda", "name": "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "password": "-", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:24:41 +0000 (0:00:00.237) 0:02:14.745 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:173 Thursday 23 November 2023 00:24:42 +0000 (0:00:00.834) 0:02:15.580 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:24:42 +0000 (0:00:00.041) 0:02:15.621 ***** skipping: [sut] => {} TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:24:42 +0000 (0:00:00.017) 0:02:15.638 ***** ok: [sut] => { "_storage_volumes_list": [ { "_device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "_raw_device": "/dev/sda", "_raw_kernel_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "present", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:24:42 +0000 (0:00:00.019) 0:02:15.658 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "size": "10G", "type": "crypt", "uuid": "60ebfbd0-5cee-4bc0-9f11-ab2de2e0d8d7" }, "/dev/sda": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "edb67d9b-a99b-4d45-a051-d3ac5299d7cf" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.219) 0:02:15.878 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003376", "end": "2023-11-23 00:24:43.303930", "rc": 0, "start": "2023-11-23 00:24:43.300554" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.217) 0:02:16.095 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003383", "end": "2023-11-23 00:24:43.520111", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:24:43.516728" } STDOUT: luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf /dev/sda - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.215) 0:02:16.311 ***** TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.324 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.029) 0:02:16.353 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.017) 0:02:16.371 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.088) 0:02:16.460 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.017) 0:02:16.477 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2574562, "block_size": 4096, "block_total": 2600960, "block_used": 26398, "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fstype": "xfs", "inode_available": 5234685, "inode_total": 5234688, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10545405952, "size_total": 10653532160, "uuid": "60ebfbd0-5cee-4bc0-9f11-ab2de2e0d8d7" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2574562, "block_size": 4096, "block_total": 2600960, "block_used": 26398, "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fstype": "xfs", "inode_available": 5234685, "inode_total": 5234688, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10545405952, "size_total": 10653532160, "uuid": "60ebfbd0-5cee-4bc0-9f11-ab2de2e0d8d7" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.022) 0:02:16.500 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.514 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.018) 0:02:16.533 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.018) 0:02:16.551 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.565 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.579 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.593 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.018) 0:02:16.611 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.625 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.015) 0:02:16.640 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.013) 0:02:16.654 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.012) 0:02:16.667 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.030) 0:02:16.697 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.017) 0:02:16.715 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.017) 0:02:16.732 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.016) 0:02:16.748 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:24:43 +0000 (0:00:00.014) 0:02:16.763 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.018) 0:02:16.781 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.019) 0:02:16.801 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699081.9439902, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699078.5471199, "dev": 5, "device_type": 2048, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 574, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700699078.5471199, "nlink": 1, "path": "/dev/sda", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.217) 0:02:17.019 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.018) 0:02:17.037 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.013) 0:02:17.051 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.016) 0:02:17.067 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "disk" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.016) 0:02:17.083 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.013) 0:02:17.097 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.015) 0:02:17.113 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699081.94599, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699078.7691114, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 962, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699078.7691114, "nlink": 1, "path": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:24:44 +0000 (0:00:00.222) 0:02:17.335 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:24:47 +0000 (0:00:02.559) 0:02:19.895 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/sda" ], "delta": "0:00:00.008282", "end": "2023-11-23 00:24:47.326163", "rc": 0, "start": "2023-11-23 00:24:47.317881" } STDOUT: LUKS header information Version: 2 Epoch: 3 Metadata area: 16384 [bytes] Keyslots area: 16744448 [bytes] UUID: edb67d9b-a99b-4d45-a051-d3ac5299d7cf Label: (no label) Subsystem: (no subsystem) Flags: (no flags) Data segments: 0: crypt offset: 16777216 [bytes] length: (whole device) cipher: aes-xts-plain64 sector: 512 [bytes] Keyslots: 0: luks2 Key: 512 bits Priority: normal Cipher: aes-xts-plain64 Cipher key: 512 bits PBKDF: argon2id Time cost: 4 Memory: 717096 Threads: 2 Salt: 2f a7 6a 72 84 60 7a a8 84 6a d8 ac 6e 38 01 90 38 58 28 74 c3 51 12 f6 4c b3 3b c0 cd 55 34 2d AF stripes: 4000 AF hash: sha256 Area offset:32768 [bytes] Area length:258048 [bytes] Digest ID: 0 Tokens: Digests: 0: pbkdf2 Hash: sha256 Iterations: 92564 Salt: 43 f3 97 11 6e a9 14 8e 62 e5 81 e3 29 71 ef 3d 6c d1 99 49 8e 45 65 3f 6d f3 d8 56 ec 26 91 d0 Digest: d9 a0 bd 9e 51 b3 a2 de d4 6a e4 b0 37 d8 9a e2 0f 48 95 bd 2b 0d 74 f1 01 be a6 04 61 6b c9 57 TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.221) 0:02:20.117 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.020) 0:02:20.137 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.021) 0:02:20.158 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.018) 0:02:20.177 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.017) 0:02:20.195 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.210 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.014) 0:02:20.225 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.016) 0:02:20.241 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf /dev/sda -" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.019) 0:02:20.260 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.018) 0:02:20.278 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.020) 0:02:20.299 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.024) 0:02:20.323 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.024) 0:02:20.348 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.017) 0:02:20.366 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.020) 0:02:20.387 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.016) 0:02:20.404 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.419 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.019) 0:02:20.438 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.014) 0:02:20.453 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.020) 0:02:20.474 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.038) 0:02:20.512 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.014) 0:02:20.526 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.541 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.014) 0:02:20.556 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.016) 0:02:20.572 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.587 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.017) 0:02:20.605 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.019) 0:02:20.624 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.019) 0:02:20.644 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.659 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.014) 0:02:20.674 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.689 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.016) 0:02:20.706 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.013) 0:02:20.720 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.735 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:24:47 +0000 (0:00:00.015) 0:02:20.751 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.016) 0:02:20.768 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:20.782 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.015) 0:02:20.798 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:20.812 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.013) 0:02:20.826 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:20.840 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:20.855 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:20.869 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.017) 0:02:20.886 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.018) 0:02:20.905 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.017) 0:02:20.922 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.017) 0:02:20.940 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.018) 0:02:20.958 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.016) 0:02:20.974 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.018) 0:02:20.992 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.015) 0:02:21.008 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:21.023 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:21.037 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:21.052 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:21.066 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.017) 0:02:21.084 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:21.098 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.013) 0:02:21.111 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Test for correct handling of new encrypted volume w/ no key] ************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:180 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.012) 0:02:21.124 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.035) 0:02:21.160 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.018) 0:02:21.178 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.020) 0:02:21.199 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.021) 0:02:21.221 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.016) 0:02:21.237 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.038) 0:02:21.276 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.014) 0:02:21.290 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.015) 0:02:21.306 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.013) 0:02:21.319 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.013) 0:02:21.333 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:24:48 +0000 (0:00:00.035) 0:02:21.368 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:24:51 +0000 (0:00:02.518) 0:02:23.886 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "partition", "volumes": [ { "encryption": true, "mount_point": "/opt/test1", "name": "test1", "size": "4g", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:24:51 +0000 (0:00:00.020) 0:02:23.907 ***** ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:24:51 +0000 (0:00:00.018) 0:02:23.926 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:24:52 +0000 (0:00:01.753) 0:02:25.679 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:24:52 +0000 (0:00:00.028) 0:02:25.708 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:24:52 +0000 (0:00:00.023) 0:02:25.732 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:24:52 +0000 (0:00:00.014) 0:02:25.746 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:24:53 +0000 (0:00:00.022) 0:02:25.769 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:24:55 +0000 (0:00:02.495) 0:02:28.265 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:24:57 +0000 (0:00:02.415) 0:02:30.680 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:24:57 +0000 (0:00:00.025) 0:02:30.706 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:24:57 +0000 (0:00:00.013) 0:02:30.720 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: encrypted volume 'test1' missing key/password TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:24:59 +0000 (0:00:01.979) 0:02:32.699 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "encrypted volume 'test1' missing key/password", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:24:59 +0000 (0:00:00.019) 0:02:32.719 ***** TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:24:59 +0000 (0:00:00.013) 0:02:32.733 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:24:59 +0000 (0:00:00.016) 0:02:32.749 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.023) 0:02:32.773 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Create an encrypted partition volume w/ default fs] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:199 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.014) 0:02:32.788 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.055) 0:02:32.844 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.021) 0:02:32.866 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.017) 0:02:32.883 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.093) 0:02:32.977 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.016) 0:02:32.993 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.016) 0:02:33.009 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.014) 0:02:33.023 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.013) 0:02:33.037 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:25:00 +0000 (0:00:00.030) 0:02:33.068 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:25:02 +0000 (0:00:02.533) 0:02:35.602 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "partition", "volumes": [ { "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:25:02 +0000 (0:00:00.018) 0:02:35.621 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:25:02 +0000 (0:00:00.015) 0:02:35.636 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:25:04 +0000 (0:00:01.736) 0:02:37.372 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:25:04 +0000 (0:00:00.028) 0:02:37.401 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:25:04 +0000 (0:00:00.024) 0:02:37.426 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:25:04 +0000 (0:00:00.015) 0:02:37.441 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:25:04 +0000 (0:00:00.023) 0:02:37.465 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:25:07 +0000 (0:00:02.504) 0:02:39.969 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:25:09 +0000 (0:00:02.440) 0:02:42.410 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:25:09 +0000 (0:00:00.025) 0:02:42.436 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:25:09 +0000 (0:00:00.013) 0:02:42.449 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "password": "-", "state": "absent" }, { "backing_device": "/dev/sda1", "name": "luks-9a411414-453f-4ae1-a354-cc11fc20f601", "password": "-", "state": "present" } ], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "mounted" } ], "packages": [ "xfsprogs", "e2fsprogs", "cryptsetup" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:25:21 +0000 (0:00:11.620) 0:02:54.069 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:25:21 +0000 (0:00:00.016) 0:02:54.086 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:25:21 +0000 (0:00:00.013) 0:02:54.099 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "luks" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda", "name": "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "password": "-", "state": "absent" }, { "backing_device": "/dev/sda1", "name": "luks-9a411414-453f-4ae1-a354-cc11fc20f601", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "mounted" } ], "packages": [ "xfsprogs", "e2fsprogs", "cryptsetup" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:25:21 +0000 (0:00:00.020) 0:02:54.119 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:25:21 +0000 (0:00:00.022) 0:02:54.142 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:25:21 +0000 (0:00:00.017) 0:02:54.159 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:25:21 +0000 (0:00:00.234) 0:02:54.394 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:25:22 +0000 (0:00:01.023) 0:02:55.417 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:25:22 +0000 (0:00:00.251) 0:02:55.669 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:25:22 +0000 (0:00:00.020) 0:02:55.689 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:25:23 +0000 (0:00:01.007) 0:02:56.696 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699081.9419901, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "b4a213e629f9a3d929602807548af3ed900e50d4", "ctime": 1700699081.9379904, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699081.9369905, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 53, "uid": 0, "version": "1410139664", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:25:24 +0000 (0:00:00.225) 0:02:56.921 ***** changed: [sut] => (item={'backing_device': '/dev/sda', 'name': 'luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf', 'password': '-', 'state': 'absent'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda", "name": "luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed changed: [sut] => (item={'backing_device': '/dev/sda1', 'name': 'luks-9a411414-453f-4ae1-a354-cc11fc20f601', 'password': '-', 'state': 'present'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda1", "name": "luks-9a411414-453f-4ae1-a354-cc11fc20f601", "password": "-", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:25:24 +0000 (0:00:00.470) 0:02:57.392 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:216 Thursday 23 November 2023 00:25:25 +0000 (0:00:00.820) 0:02:58.213 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:25:25 +0000 (0:00:00.042) 0:02:58.255 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:25:25 +0000 (0:00:00.022) 0:02:58.277 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:25:25 +0000 (0:00:00.021) 0:02:58.299 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "size": "10G", "type": "crypt", "uuid": "7894f2fa-3b67-4d11-91e1-8e223aa98f61" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "9a411414-453f-4ae1-a354-cc11fc20f601" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:25:25 +0000 (0:00:00.225) 0:02:58.524 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003439", "end": "2023-11-23 00:25:25.949943", "rc": 0, "start": "2023-11-23 00:25:25.946504" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:25:25 +0000 (0:00:00.216) 0:02:58.741 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003463", "end": "2023-11-23 00:25:26.163672", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:25:26.160209" } STDOUT: luks-9a411414-453f-4ae1-a354-cc11fc20f601 /dev/sda1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.212) 0:02:58.954 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.031) 0:02:58.986 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.000 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.033) 0:02:59.034 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.048 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.015) 0:02:59.064 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.079 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.015) 0:02:59.094 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.108 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.123 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.137 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.151 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.166 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.051) 0:02:59.218 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.015) 0:02:59.234 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.249 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.013) 0:02:59.263 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.278 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.013) 0:02:59.291 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.013) 0:02:59.305 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.013) 0:02:59.319 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.013) 0:02:59.332 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.346 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.016) 0:02:59.362 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.377 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.033) 0:02:59.411 ***** skipping: [sut] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_lvmraid_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_lvmraid_volume": { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.027) 0:02:59.439 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.034) 0:02:59.474 ***** skipping: [sut] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_thin_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_thin_volume": { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.021) 0:02:59.495 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.034) 0:02:59.530 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.017) 0:02:59.548 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.015) 0:02:59.563 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.015) 0:02:59.579 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.593 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.036) 0:02:59.630 ***** skipping: [sut] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_vdo_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_vdo_volume": { "_device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.019) 0:02:59.649 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.014) 0:02:59.663 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.029) 0:02:59.692 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:25:26 +0000 (0:00:00.019) 0:02:59.712 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.071) 0:02:59.784 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.017) 0:02:59.802 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2572529, "block_size": 4096, "block_total": 2598912, "block_used": 26383, "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fstype": "xfs", "inode_available": 5230589, "inode_total": 5230592, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10537078784, "size_total": 10645143552, "uuid": "7894f2fa-3b67-4d11-91e1-8e223aa98f61" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2572529, "block_size": 4096, "block_total": 2598912, "block_used": 26383, "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fstype": "xfs", "inode_available": 5230589, "inode_total": 5230592, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10537078784, "size_total": 10645143552, "uuid": "7894f2fa-3b67-4d11-91e1-8e223aa98f61" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.022) 0:02:59.824 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:02:59.838 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.017) 0:02:59.856 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.016) 0:02:59.873 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:02:59.887 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.014) 0:02:59.901 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:02:59.915 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.017) 0:02:59.933 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:02:59.946 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:02:59.960 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:02:59.974 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.014) 0:02:59.988 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.030) 0:03:00.019 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.017) 0:03:00.036 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.017) 0:03:00.053 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:03:00.067 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.015) 0:03:00.082 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.069) 0:03:00.152 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.021) 0:03:00.173 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699124.587362, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699120.9525008, "dev": 5, "device_type": 2049, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1081, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700699120.9525008, "nlink": 1, "path": "/dev/sda1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.221) 0:03:00.395 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.020) 0:03:00.415 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.015) 0:03:00.430 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.017) 0:03:00.448 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "partition" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.014) 0:03:00.463 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.013) 0:03:00.476 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.018) 0:03:00.494 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699124.591362, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699121.1764925, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1137, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699121.1764925, "nlink": 1, "path": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:25:27 +0000 (0:00:00.219) 0:03:00.714 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:25:30 +0000 (0:00:02.507) 0:03:03.221 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/sda1" ], "delta": "0:00:00.008275", "end": "2023-11-23 00:25:30.652457", "rc": 0, "start": "2023-11-23 00:25:30.644182" } STDOUT: LUKS header information Version: 2 Epoch: 3 Metadata area: 16384 [bytes] Keyslots area: 16744448 [bytes] UUID: 9a411414-453f-4ae1-a354-cc11fc20f601 Label: (no label) Subsystem: (no subsystem) Flags: (no flags) Data segments: 0: crypt offset: 16777216 [bytes] length: (whole device) cipher: aes-xts-plain64 sector: 512 [bytes] Keyslots: 0: luks2 Key: 512 bits Priority: normal Cipher: aes-xts-plain64 Cipher key: 512 bits PBKDF: argon2id Time cost: 4 Memory: 708496 Threads: 2 Salt: 51 36 30 9d 59 20 6d 1b cd 7a 5e 88 55 0b fe 19 a5 39 73 a2 75 a0 6e 92 3b 91 1d a1 0b f5 59 e8 AF stripes: 4000 AF hash: sha256 Area offset:32768 [bytes] Area length:258048 [bytes] Digest ID: 0 Tokens: Digests: 0: pbkdf2 Hash: sha256 Iterations: 93223 Salt: b6 c7 b6 64 03 bc 55 f4 88 fa 25 84 43 44 fe 77 2b c2 d9 ef 2c 85 9d 48 1b 43 a3 5a 24 79 9c 5b Digest: 56 54 8c 9c 02 fd 52 32 58 bf e1 2c 20 21 b9 08 1c ce 5a a7 79 d1 4b ce 39 88 e3 f2 3b c8 5c 60 TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.221) 0:03:03.443 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.020) 0:03:03.463 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.020) 0:03:03.484 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.018) 0:03:03.503 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.018) 0:03:03.521 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.015) 0:03:03.536 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.015) 0:03:03.552 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.015) 0:03:03.567 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-9a411414-453f-4ae1-a354-cc11fc20f601 /dev/sda1 -" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.018) 0:03:03.586 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.017) 0:03:03.603 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.023) 0:03:03.627 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.018) 0:03:03.645 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.018) 0:03:03.664 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.013) 0:03:03.678 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.013) 0:03:03.692 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.014) 0:03:03.706 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.015) 0:03:03.721 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.014) 0:03:03.735 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.013) 0:03:03.749 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:25:30 +0000 (0:00:00.014) 0:03:03.763 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:03.777 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:03.791 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.806 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:03.820 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.835 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.850 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.866 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.881 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.016) 0:03:03.898 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.913 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.928 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.944 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.959 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:03.973 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:03.989 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:04.005 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.019 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.034 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:04.050 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.064 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:04.080 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:04.094 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.108 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.122 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:04.136 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:04.150 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.015) 0:03:04.165 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.180 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.017) 0:03:04.197 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.016) 0:03:04.214 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.016) 0:03:04.231 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.245 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.016) 0:03:04.261 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.276 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.291 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.305 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.319 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.334 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.039) 0:03:04.373 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.013) 0:03:04.387 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml:12 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.014) 0:03:04.401 ***** changed: [sut] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Test for correct handling of safe_mode] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:222 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.224) 0:03:04.626 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.039) 0:03:04.665 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.019) 0:03:04.685 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.020) 0:03:04.705 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.021) 0:03:04.727 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:25:31 +0000 (0:00:00.018) 0:03:04.746 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:25:32 +0000 (0:00:00.036) 0:03:04.782 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:25:32 +0000 (0:00:00.014) 0:03:04.797 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:25:32 +0000 (0:00:00.014) 0:03:04.811 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:25:32 +0000 (0:00:00.013) 0:03:04.824 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:25:32 +0000 (0:00:00.013) 0:03:04.837 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:25:32 +0000 (0:00:00.032) 0:03:04.869 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:25:34 +0000 (0:00:02.496) 0:03:07.366 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "partition", "volumes": [ { "encryption": false, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:25:34 +0000 (0:00:00.020) 0:03:07.386 ***** ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:25:34 +0000 (0:00:00.018) 0:03:07.405 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:25:36 +0000 (0:00:01.956) 0:03:09.362 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:25:36 +0000 (0:00:00.029) 0:03:09.391 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:25:36 +0000 (0:00:00.024) 0:03:09.415 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:25:36 +0000 (0:00:00.015) 0:03:09.431 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:25:36 +0000 (0:00:00.022) 0:03:09.453 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:25:39 +0000 (0:00:02.589) 0:03:12.042 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service": { "name": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:25:41 +0000 (0:00:02.450) 0:03:14.493 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:25:41 +0000 (0:00:00.025) 0:03:14.518 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2dedb67d9b\x2da99b\x2d4d45\x2da051\x2dd3ac5299d7cf.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "name": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-udevd-kernel.socket systemd-journald.socket dev-sda.device \"system-systemd\\\\x2dcryptsetup.slice\" cryptsetup-pre.target", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "umount.target \"blockdev@dev-mapper-luks\\\\x2dedb67d9b\\\\x2da99b\\\\x2d4d45\\\\x2da051\\\\x2dd3ac5299d7cf.target\" cryptsetup.target", "BindsTo": "dev-sda.device", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf /dev/sda - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf /dev/sda - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-edb67d9b-a99b-4d45-a051-d3ac5299d7cf ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2dedb67d9b\\\\x2da99b\\\\x2d4d45\\\\x2da051\\\\x2dd3ac5299d7cf.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:25:23 UTC", "StateChangeTimestampMonotonic": "1058790895", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2dedb67d9b\\\\x2da99b\\\\x2d4d45\\\\x2da051\\\\x2dd3ac5299d7cf.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:25:42 +0000 (0:00:01.053) 0:03:15.572 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting on device 'luks-9a411414-453f-4ae1-a354-cc11fc20f601' in safe mode due to encryption removal TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:25:44 +0000 (0:00:01.979) 0:03:17.551 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting on device 'luks-9a411414-453f-4ae1-a354-cc11fc20f601' in safe mode due to encryption removal", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:25:44 +0000 (0:00:00.019) 0:03:17.571 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2dedb67d9b\x2da99b\x2d4d45\x2da051\x2dd3ac5299d7cf.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "name": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2dedb67d9b\\x2da99b\\x2d4d45\\x2da051\\x2dd3ac5299d7cf.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2dedb67d9b\\\\x2da99b\\\\x2d4d45\\\\x2da051\\\\x2dd3ac5299d7cf.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:25:45 +0000 (0:00:01.043) 0:03:18.615 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:25:45 +0000 (0:00:00.016) 0:03:18.631 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:25:45 +0000 (0:00:00.021) 0:03:18.653 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:11 Thursday 23 November 2023 00:25:45 +0000 (0:00:00.014) 0:03:18.667 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699131.8340855, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699131.8340855, "dev": 64768, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1700699131.8340855, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "915802217", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Assert file presence] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:16 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.217) 0:03:18.884 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Remove the encryption layer] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:246 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.018) 0:03:18.902 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.061) 0:03:18.964 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.021) 0:03:18.985 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.015) 0:03:19.001 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.036) 0:03:19.038 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.014) 0:03:19.052 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.013) 0:03:19.066 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.012) 0:03:19.078 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.013) 0:03:19.092 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:25:46 +0000 (0:00:00.054) 0:03:19.147 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:25:48 +0000 (0:00:02.512) 0:03:21.660 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "partition", "volumes": [ { "encryption": false, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:25:48 +0000 (0:00:00.018) 0:03:21.678 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:25:48 +0000 (0:00:00.016) 0:03:21.695 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:25:50 +0000 (0:00:01.943) 0:03:23.639 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:25:50 +0000 (0:00:00.027) 0:03:23.667 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:25:50 +0000 (0:00:00.024) 0:03:23.691 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:25:50 +0000 (0:00:00.015) 0:03:23.706 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:25:50 +0000 (0:00:00.022) 0:03:23.728 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:25:53 +0000 (0:00:02.518) 0:03:26.247 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service": { "name": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:25:55 +0000 (0:00:02.435) 0:03:28.682 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:25:55 +0000 (0:00:00.025) 0:03:28.707 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d9a411414\x2d453f\x2d4ae1\x2da354\x2dcc11fc20f601.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "name": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "dev-sda1.device systemd-journald.socket cryptsetup-pre.target \"system-systemd\\\\x2dcryptsetup.slice\" systemd-udevd-kernel.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "cryptsetup.target umount.target \"blockdev@dev-mapper-luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.target\"", "BindsTo": "dev-sda1.device", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-9a411414-453f-4ae1-a354-cc11fc20f601", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-9a411414-453f-4ae1-a354-cc11fc20f601 /dev/sda1 - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-9a411414-453f-4ae1-a354-cc11fc20f601 /dev/sda1 - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-9a411414-453f-4ae1-a354-cc11fc20f601 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-9a411414-453f-4ae1-a354-cc11fc20f601 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.service\"", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target \"dev-mapper-luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.device\"", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:25:45 UTC", "StateChangeTimestampMonotonic": "1080690119", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:25:56 +0000 (0:00:01.050) 0:03:29.758 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/sda1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda1", "name": "luks-9a411414-453f-4ae1-a354-cc11fc20f601", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sda1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:25:59 +0000 (0:00:02.692) 0:03:32.450 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:25:59 +0000 (0:00:00.016) 0:03:32.466 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d9a411414\x2d453f\x2d4ae1\x2da354\x2dcc11fc20f601.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "name": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "RequiredBy": "\"dev-mapper-luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.device\" cryptsetup.target", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:25:45 UTC", "StateChangeTimestampMonotonic": "1080690119", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:26:00 +0000 (0:00:01.045) 0:03:33.511 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/sda1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda1", "name": "luks-9a411414-453f-4ae1-a354-cc11fc20f601", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sda1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:26:00 +0000 (0:00:00.018) 0:03:33.530 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:26:00 +0000 (0:00:00.018) 0:03:33.548 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:26:00 +0000 (0:00:00.016) 0:03:33.564 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-9a411414-453f-4ae1-a354-cc11fc20f601" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:26:01 +0000 (0:00:00.232) 0:03:33.797 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:26:02 +0000 (0:00:01.019) 0:03:34.817 ***** changed: [sut] => (item={'src': 'UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:26:02 +0000 (0:00:00.286) 0:03:35.103 ***** skipping: [sut] => (item={'src': 'UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:26:02 +0000 (0:00:00.023) 0:03:35.126 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:26:03 +0000 (0:00:01.025) 0:03:36.152 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699124.5853622, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "4c099542a27045bc035931d504b5495fa7cd325a", "ctime": 1700699124.5833623, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699124.5833623, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 54, "uid": 0, "version": "2494225383", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:26:03 +0000 (0:00:00.232) 0:03:36.385 ***** changed: [sut] => (item={'backing_device': '/dev/sda1', 'name': 'luks-9a411414-453f-4ae1-a354-cc11fc20f601', 'password': '-', 'state': 'absent'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda1", "name": "luks-9a411414-453f-4ae1-a354-cc11fc20f601", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:26:03 +0000 (0:00:00.241) 0:03:36.626 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:263 Thursday 23 November 2023 00:26:04 +0000 (0:00:00.820) 0:03:37.446 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:26:04 +0000 (0:00:00.040) 0:03:37.487 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:26:04 +0000 (0:00:00.020) 0:03:37.508 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:26:04 +0000 (0:00:00.013) 0:03:37.521 ***** ok: [sut] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "xfs", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "81cf8a9c-a044-43c7-9b3f-6a0a1629240a" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:26:04 +0000 (0:00:00.220) 0:03:37.742 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003434", "end": "2023-11-23 00:26:05.165842", "rc": 0, "start": "2023-11-23 00:26:05.162408" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.214) 0:03:37.956 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003298", "end": "2023-11-23 00:26:05.380707", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:26:05.377409" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.215) 0:03:38.172 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.031) 0:03:38.204 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.220 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.032) 0:03:38.253 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.270 ***** TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.013) 0:03:38.283 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.299 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.040) 0:03:38.340 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.356 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.372 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.388 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.018) 0:03:38.407 ***** TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.423 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.040) 0:03:38.463 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.480 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.496 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.512 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.528 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.543 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.560 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.015) 0:03:38.576 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.592 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.019) 0:03:38.612 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.016) 0:03:38.628 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.014) 0:03:38.643 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.030) 0:03:38.674 ***** skipping: [sut] => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/sda1', '_raw_device': '/dev/sda1', '_mount_id': 'UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a', '_kernel_device': '/dev/sda1', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_lvmraid_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_lvmraid_volume": { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.019) 0:03:38.693 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.031) 0:03:38.725 ***** skipping: [sut] => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/sda1', '_raw_device': '/dev/sda1', '_mount_id': 'UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a', '_kernel_device': '/dev/sda1', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_thin_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_thin_volume": { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:26:05 +0000 (0:00:00.019) 0:03:38.745 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.036) 0:03:38.782 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.019) 0:03:38.801 ***** TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.017) 0:03:38.819 ***** TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.013) 0:03:38.833 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.013) 0:03:38.846 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.033) 0:03:38.880 ***** skipping: [sut] => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks2', 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/sda1', '_raw_device': '/dev/sda1', '_mount_id': 'UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a', '_kernel_device': '/dev/sda1', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_vdo_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_vdo_volume": { "_device": "/dev/sda1", "_kernel_device": "/dev/sda1", "_mount_id": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks2", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.019) 0:03:38.899 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.014) 0:03:38.914 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.028) 0:03:38.942 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.020) 0:03:38.962 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.086) 0:03:39.049 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/sda1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.018) 0:03:39.067 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2576597, "block_size": 4096, "block_total": 2603008, "block_used": 26411, "device": "/dev/sda1", "fstype": "xfs", "inode_available": 5238781, "inode_total": 5238784, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10553741312, "size_total": 10661920768, "uuid": "81cf8a9c-a044-43c7-9b3f-6a0a1629240a" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2576597, "block_size": 4096, "block_total": 2603008, "block_used": 26411, "device": "/dev/sda1", "fstype": "xfs", "inode_available": 5238781, "inode_total": 5238784, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10553741312, "size_total": 10661920768, "uuid": "81cf8a9c-a044-43c7-9b3f-6a0a1629240a" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.024) 0:03:39.091 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.017) 0:03:39.109 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.021) 0:03:39.131 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.022) 0:03:39.153 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.019) 0:03:39.173 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.016) 0:03:39.189 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.014) 0:03:39.204 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.020) 0:03:39.224 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.014) 0:03:39.239 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.014) 0:03:39.253 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.014) 0:03:39.267 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.014) 0:03:39.281 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.037) 0:03:39.318 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.052) 0:03:39.371 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.020) 0:03:39.391 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.017) 0:03:39.409 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.016) 0:03:39.425 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.022) 0:03:39.448 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.020) 0:03:39.469 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699163.8198643, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699159.5720265, "dev": 5, "device_type": 2049, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1332, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700699159.5720265, "nlink": 1, "path": "/dev/sda1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.225) 0:03:39.694 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.022) 0:03:39.716 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.019) 0:03:39.736 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:26:06 +0000 (0:00:00.020) 0:03:39.756 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "partition" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:26:07 +0000 (0:00:00.017) 0:03:39.773 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:26:07 +0000 (0:00:00.015) 0:03:39.789 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:26:07 +0000 (0:00:00.017) 0:03:39.807 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:26:07 +0000 (0:00:00.014) 0:03:39.821 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:26:09 +0000 (0:00:02.620) 0:03:42.442 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.017) 0:03:42.459 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.015) 0:03:42.475 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.021) 0:03:42.496 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.510 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.524 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.015) 0:03:42.539 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.553 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.566 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.022) 0:03:42.589 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.018) 0:03:42.607 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.014) 0:03:42.622 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.016) 0:03:42.639 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.653 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.666 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.680 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.694 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.708 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.015) 0:03:42.723 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.013) 0:03:42.737 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:26:09 +0000 (0:00:00.014) 0:03:42.752 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.768 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:42.782 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:42.797 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.016) 0:03:42.813 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.828 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.843 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.858 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.874 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.889 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.016) 0:03:42.906 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.921 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.936 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.951 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:42.965 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:42.979 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:42.994 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.008 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.022 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.036 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.050 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.064 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:43.079 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.093 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.106 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.121 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.135 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.150 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:43.165 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.038) 0:03:43.204 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.015) 0:03:43.220 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.235 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.248 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.262 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.275 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.014) 0:03:43.290 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.304 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.318 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.331 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.012) 0:03:43.344 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.012) 0:03:43.356 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml:12 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.013) 0:03:43.370 ***** changed: [sut] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Test for correct handling of safe_mode] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:269 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.220) 0:03:43.591 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.044) 0:03:43.635 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.018) 0:03:43.654 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.020) 0:03:43.674 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.020) 0:03:43.695 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.016) 0:03:43.712 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:26:10 +0000 (0:00:00.039) 0:03:43.751 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:26:11 +0000 (0:00:00.014) 0:03:43.766 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:26:11 +0000 (0:00:00.014) 0:03:43.781 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:26:11 +0000 (0:00:00.014) 0:03:43.796 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:26:11 +0000 (0:00:00.014) 0:03:43.810 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:26:11 +0000 (0:00:00.033) 0:03:43.843 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:26:13 +0000 (0:00:02.488) 0:03:46.331 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "partition", "volumes": [ { "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:26:13 +0000 (0:00:00.019) 0:03:46.351 ***** ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:26:13 +0000 (0:00:00.018) 0:03:46.369 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:26:15 +0000 (0:00:01.778) 0:03:48.148 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:26:15 +0000 (0:00:00.028) 0:03:48.176 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:26:15 +0000 (0:00:00.023) 0:03:48.200 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:26:15 +0000 (0:00:00.014) 0:03:48.214 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:26:15 +0000 (0:00:00.023) 0:03:48.238 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:26:17 +0000 (0:00:02.503) 0:03:50.741 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service": { "name": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:26:20 +0000 (0:00:02.432) 0:03:53.173 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:26:20 +0000 (0:00:00.024) 0:03:53.197 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d9a411414\x2d453f\x2d4ae1\x2da354\x2dcc11fc20f601.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "name": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "\"system-systemd\\\\x2dcryptsetup.slice\" systemd-udevd-kernel.socket systemd-journald.socket cryptsetup-pre.target dev-sda1.device", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "\"blockdev@dev-mapper-luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.target\" cryptsetup.target umount.target", "BindsTo": "dev-sda1.device", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-9a411414-453f-4ae1-a354-cc11fc20f601", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-9a411414-453f-4ae1-a354-cc11fc20f601 /dev/sda1 - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-9a411414-453f-4ae1-a354-cc11fc20f601 /dev/sda1 - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-9a411414-453f-4ae1-a354-cc11fc20f601 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-9a411414-453f-4ae1-a354-cc11fc20f601 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:25:45 UTC", "StateChangeTimestampMonotonic": "1080690119", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:26:21 +0000 (0:00:01.047) 0:03:54.245 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting on device 'sda1' in safe mode due to adding encryption TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:26:23 +0000 (0:00:01.842) 0:03:56.087 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting on device 'sda1' in safe mode due to adding encryption", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'state': 'present', 'type': 'partition', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:26:23 +0000 (0:00:00.019) 0:03:56.107 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d9a411414\x2d453f\x2d4ae1\x2da354\x2dcc11fc20f601.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "name": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d9a411414\\x2d453f\\x2d4ae1\\x2da354\\x2dcc11fc20f601.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d9a411414\\\\x2d453f\\\\x2d4ae1\\\\x2da354\\\\x2dcc11fc20f601.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:26:24 +0000 (0:00:01.042) 0:03:57.149 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:26:24 +0000 (0:00:00.016) 0:03:57.166 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:26:24 +0000 (0:00:00.022) 0:03:57.189 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:11 Thursday 23 November 2023 00:26:24 +0000 (0:00:00.013) 0:03:57.203 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699170.8015978, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699170.8015978, "dev": 2049, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1700699170.8015978, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "1413092986", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Assert file presence] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:16 Thursday 23 November 2023 00:26:24 +0000 (0:00:00.217) 0:03:57.420 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Create a key file] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:295 Thursday 23 November 2023 00:26:24 +0000 (0:00:00.018) 0:03:57.438 ***** ok: [sut] => { "changed": false, "gid": 0, "group": "root", "mode": "0600", "owner": "root", "path": "/tmp/storage_testw40hwjimlukskey", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Write the key into the key file] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:302 Thursday 23 November 2023 00:26:24 +0000 (0:00:00.296) 0:03:57.735 ***** ok: [sut] => { "changed": false, "checksum": "7a4dff3752e2baf5617c57eaac048e2b95e8af91", "dest": "/tmp/storage_testw40hwjimlukskey", "gid": 0, "group": "root", "md5sum": "4ac07b967150835c00d0865161e48744", "mode": "0600", "owner": "root", "secontext": "unconfined_u:object_r:user_tmp_t:s0", "size": 32, "src": "/root/.ansible/tmp/ansible-tmp-1700699185.0088809-99763-250874901627621/source", "state": "file", "uid": 0 } TASK [Add encryption to the volume] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:309 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.666) 0:03:58.402 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.023) 0:03:58.425 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.022) 0:03:58.447 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.019) 0:03:58.467 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.039) 0:03:58.507 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.013) 0:03:58.521 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.014) 0:03:58.535 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.013) 0:03:58.548 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.013) 0:03:58.561 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:26:25 +0000 (0:00:00.031) 0:03:58.593 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:26:28 +0000 (0:00:02.514) 0:04:01.108 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "partition", "volumes": [ { "encryption": true, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "mount_point": "/opt/test1", "name": "test1", "size": "4g", "type": "partition" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:26:28 +0000 (0:00:00.018) 0:04:01.127 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:26:28 +0000 (0:00:00.015) 0:04:01.142 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:26:30 +0000 (0:00:01.802) 0:04:02.945 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:26:30 +0000 (0:00:00.027) 0:04:02.972 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:26:30 +0000 (0:00:00.023) 0:04:02.996 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:26:30 +0000 (0:00:00.014) 0:04:03.011 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:26:30 +0000 (0:00:00.023) 0:04:03.035 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:26:32 +0000 (0:00:02.495) 0:04:05.530 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:26:35 +0000 (0:00:02.454) 0:04:07.985 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:26:35 +0000 (0:00:00.024) 0:04:08.010 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:26:35 +0000 (0:00:00.015) 0:04:08.025 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/sda1", "fs_type": "xfs" }, { "action": "create format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda1", "name": "luks-0ff83410-148e-4f69-832c-c589e3174417", "password": "/tmp/storage_testw40hwjimlukskey", "state": "present" } ], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "mounted" } ], "packages": [ "e2fsprogs", "cryptsetup", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:26:46 +0000 (0:00:11.046) 0:04:19.071 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:26:46 +0000 (0:00:00.015) 0:04:19.087 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:26:46 +0000 (0:00:00.013) 0:04:19.101 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/sda1", "fs_type": "xfs" }, { "action": "create format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda1", "name": "luks-0ff83410-148e-4f69-832c-c589e3174417", "password": "/tmp/storage_testw40hwjimlukskey", "state": "present" } ], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "mounted" } ], "packages": [ "e2fsprogs", "cryptsetup", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:26:46 +0000 (0:00:00.018) 0:04:19.119 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:26:46 +0000 (0:00:00.019) 0:04:19.139 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:26:46 +0000 (0:00:00.016) 0:04:19.155 ***** changed: [sut] => (item={'src': 'UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "UUID=81cf8a9c-a044-43c7-9b3f-6a0a1629240a" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:26:46 +0000 (0:00:00.230) 0:04:19.385 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:26:47 +0000 (0:00:01.012) 0:04:20.398 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:26:47 +0000 (0:00:00.247) 0:04:20.646 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:26:47 +0000 (0:00:00.021) 0:04:20.667 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:26:48 +0000 (0:00:01.019) 0:04:21.687 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699165.378805, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699163.8168645, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263320, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1700699163.8158646, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "1890776070", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:26:49 +0000 (0:00:00.222) 0:04:21.909 ***** changed: [sut] => (item={'backing_device': '/dev/sda1', 'name': 'luks-0ff83410-148e-4f69-832c-c589e3174417', 'password': '/tmp/storage_testw40hwjimlukskey', 'state': 'present'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda1", "name": "luks-0ff83410-148e-4f69-832c-c589e3174417", "password": "/tmp/storage_testw40hwjimlukskey", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:26:49 +0000 (0:00:00.241) 0:04:22.151 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:326 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.831) 0:04:22.982 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.023) 0:04:23.005 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "partition", "volumes": [ { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.064) 0:04:23.070 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.014) 0:04:23.085 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "size": "10G", "type": "crypt", "uuid": "e59a3518-e9fe-4ca4-9147-4b8bd5bc86e8" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "0ff83410-148e-4f69-832c-c589e3174417" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.216) 0:04:23.301 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003343", "end": "2023-11-23 00:26:50.723550", "rc": 0, "start": "2023-11-23 00:26:50.720207" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.212) 0:04:23.513 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003322", "end": "2023-11-23 00:26:50.933959", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:26:50.930637" } STDOUT: luks-0ff83410-148e-4f69-832c-c589e3174417 /dev/sda1 /tmp/storage_testw40hwjimlukskey TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.210) 0:04:23.724 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:26:50 +0000 (0:00:00.031) 0:04:23.755 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:23.769 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.031) 0:04:23.801 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:23.816 ***** TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:23.830 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:23.844 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:23.858 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:23.873 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:23.888 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:23.904 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.017) 0:04:23.921 ***** TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:23.935 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.030) 0:04:23.965 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:23.980 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:23.996 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.010 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.023 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.038 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.052 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.065 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:24.081 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.096 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.110 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.124 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.029) 0:04:24.153 ***** skipping: [sut] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': '/tmp/storage_testw40hwjimlukskey', 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_lvmraid_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_lvmraid_volume": { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.024) 0:04:24.178 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.030) 0:04:24.209 ***** skipping: [sut] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': '/tmp/storage_testw40hwjimlukskey', 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_thin_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_thin_volume": { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.020) 0:04:24.229 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.034) 0:04:24.264 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.017) 0:04:24.281 ***** TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.294 ***** TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.012) 0:04:24.307 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:24.323 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.034) 0:04:24.357 ***** skipping: [sut] => (item={'encryption': True, 'encryption_cipher': None, 'encryption_key': '/tmp/storage_testw40hwjimlukskey', 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'partition', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', '_raw_device': '/dev/sda1', '_mount_id': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', '_kernel_device': '/dev/dm-0', '_raw_kernel_device': '/dev/sda1'}) => { "ansible_loop_var": "storage_test_vdo_volume", "changed": false, "skip_reason": "Conditional result was False", "storage_test_vdo_volume": { "_device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "_raw_device": "/dev/sda1", "_raw_kernel_device": "/dev/sda1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": "/tmp/storage_testw40hwjimlukskey", "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "partition", "vdo_pool_size": null } } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.019) 0:04:24.376 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.390 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.052) 0:04:24.443 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.019) 0:04:24.462 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.071) 0:04:24.534 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.017) 0:04:24.551 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 2572529, "block_size": 4096, "block_total": 2598912, "block_used": 26383, "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fstype": "xfs", "inode_available": 5230589, "inode_total": 5230592, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10537078784, "size_total": 10645143552, "uuid": "e59a3518-e9fe-4ca4-9147-4b8bd5bc86e8" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 2572529, "block_size": 4096, "block_total": 2598912, "block_used": 26383, "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fstype": "xfs", "inode_available": 5230589, "inode_total": 5230592, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 10537078784, "size_total": 10645143552, "uuid": "e59a3518-e9fe-4ca4-9147-4b8bd5bc86e8" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.023) 0:04:24.575 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.015) 0:04:24.590 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.018) 0:04:24.608 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.018) 0:04:24.626 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.016) 0:04:24.643 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.657 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.013) 0:04:24.671 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.018) 0:04:24.690 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.704 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.719 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.016) 0:04:24.735 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:26:51 +0000 (0:00:00.014) 0:04:24.749 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.030) 0:04:24.780 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.017) 0:04:24.798 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.017) 0:04:24.815 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.015) 0:04:24.831 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.014) 0:04:24.845 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.020) 0:04:24.866 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.019) 0:04:24.885 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699209.3471243, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699205.9562545, "dev": 5, "device_type": 2049, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1332, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700699205.9562545, "nlink": 1, "path": "/dev/sda1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.217) 0:04:25.103 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.019) 0:04:25.122 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.016) 0:04:25.139 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.017) 0:04:25.157 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "partition" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.015) 0:04:25.172 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.015) 0:04:25.187 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.018) 0:04:25.206 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699209.3511243, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699206.175246, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1498, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699206.175246, "nlink": 1, "path": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:26:52 +0000 (0:00:00.221) 0:04:25.427 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:26:55 +0000 (0:00:02.527) 0:04:27.954 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/sda1" ], "delta": "0:00:00.008493", "end": "2023-11-23 00:26:55.387022", "rc": 0, "start": "2023-11-23 00:26:55.378529" } STDOUT: LUKS header information Version: 2 Epoch: 3 Metadata area: 16384 [bytes] Keyslots area: 16744448 [bytes] UUID: 0ff83410-148e-4f69-832c-c589e3174417 Label: (no label) Subsystem: (no subsystem) Flags: (no flags) Data segments: 0: crypt offset: 16777216 [bytes] length: (whole device) cipher: aes-xts-plain64 sector: 512 [bytes] Keyslots: 0: luks2 Key: 512 bits Priority: normal Cipher: aes-xts-plain64 Cipher key: 512 bits PBKDF: argon2id Time cost: 4 Memory: 708496 Threads: 2 Salt: b1 27 da d2 4d e9 4a cf 25 d2 fe 17 15 13 17 9b b4 55 d8 18 46 21 3d ee 7b ae 53 56 db 29 8e 9b AF stripes: 4000 AF hash: sha256 Area offset:32768 [bytes] Area length:258048 [bytes] Digest ID: 0 Tokens: Digests: 0: pbkdf2 Hash: sha256 Iterations: 93090 Salt: 01 a8 7e 31 61 cf 13 e1 d7 86 9a 31 60 05 c6 f2 00 e2 04 2e 85 95 ba a3 73 ed 78 cf be 32 c1 1f Digest: 08 6f 9d 47 d2 ba 0d 9f c5 e2 54 36 4a f1 eb 05 73 23 ba 70 1c 94 6e 5d b4 4f 23 cb f9 0e c7 f8 TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.224) 0:04:28.179 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.020) 0:04:28.199 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.022) 0:04:28.222 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.020) 0:04:28.243 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.024) 0:04:28.268 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.019) 0:04:28.287 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.302 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.316 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-0ff83410-148e-4f69-832c-c589e3174417 /dev/sda1 /tmp/storage_testw40hwjimlukskey" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "/tmp/storage_testw40hwjimlukskey" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.020) 0:04:28.337 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.042) 0:04:28.380 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.020) 0:04:28.400 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.018) 0:04:28.419 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.020) 0:04:28.439 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.453 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.468 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.484 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.498 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.512 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.526 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.540 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.555 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.016) 0:04:28.571 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.013) 0:04:28.585 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.013) 0:04:28.599 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.014) 0:04:28.614 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.629 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.645 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.016) 0:04:28.662 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.677 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.692 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.708 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.723 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.015) 0:04:28.738 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:26:55 +0000 (0:00:00.016) 0:04:28.754 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.769 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.783 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.797 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.013) 0:04:28.811 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.825 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:28.841 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.855 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.869 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.884 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.898 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.912 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.016) 0:04:28.928 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.943 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:28.957 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:28.972 ***** ok: [sut] => { "storage_test_expected_size": "VARIABLE IS NOT DEFINED!: 'storage_test_expected_size' is undefined" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:28.988 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:29.003 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:29.019 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.013) 0:04:29.033 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:29.048 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.063 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.077 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.092 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.015) 0:04:29.107 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.122 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.013) 0:04:29.135 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove the key file] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:329 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.012) 0:04:29.148 ***** ok: [sut] => { "changed": false, "path": "/tmp/storage_testw40hwjimlukskey", "state": "absent" } TASK [Test for correct handling of new encrypted volume w/ no key] ************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:339 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.219) 0:04:29.367 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.023) 0:04:29.391 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.017) 0:04:29.408 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.021) 0:04:29.430 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.021) 0:04:29.452 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.016) 0:04:29.468 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.038) 0:04:29.506 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.521 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.535 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.014) 0:04:29.549 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.013) 0:04:29.563 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:26:56 +0000 (0:00:00.032) 0:04:29.595 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:26:59 +0000 (0:00:02.506) 0:04:32.102 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "encryption": true, "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:26:59 +0000 (0:00:00.020) 0:04:32.122 ***** ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:26:59 +0000 (0:00:00.019) 0:04:32.141 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup", "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:27:01 +0000 (0:00:01.973) 0:04:34.114 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:27:01 +0000 (0:00:00.028) 0:04:34.143 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:27:01 +0000 (0:00:00.023) 0:04:34.167 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:27:01 +0000 (0:00:00.015) 0:04:34.182 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:27:01 +0000 (0:00:00.022) 0:04:34.205 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:27:03 +0000 (0:00:02.557) 0:04:36.763 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:27:06 +0000 (0:00:02.433) 0:04:39.197 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:27:06 +0000 (0:00:00.025) 0:04:39.223 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:27:06 +0000 (0:00:00.013) 0:04:39.236 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: encrypted volume 'test1' missing key/password TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:27:08 +0000 (0:00:02.162) 0:04:41.399 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "encrypted volume 'test1' missing key/password", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': False, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.019) 0:04:41.418 ***** TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.012) 0:04:41.431 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.016) 0:04:41.447 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.023) 0:04:41.470 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Create an encrypted lvm volume w/ default fs] **************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:357 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.013) 0:04:41.484 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.029) 0:04:41.514 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.022) 0:04:41.536 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.017) 0:04:41.553 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.037) 0:04:41.591 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.014) 0:04:41.605 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.015) 0:04:41.620 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.014) 0:04:41.635 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.015) 0:04:41.650 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:27:08 +0000 (0:00:00.033) 0:04:41.684 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:27:11 +0000 (0:00:02.564) 0:04:44.248 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "encryption": true, "encryption_cipher": "aes-xts-plain64", "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:27:11 +0000 (0:00:00.022) 0:04:44.271 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:27:11 +0000 (0:00:00.016) 0:04:44.288 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup", "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:27:13 +0000 (0:00:01.959) 0:04:46.247 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:27:13 +0000 (0:00:00.027) 0:04:46.274 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:27:13 +0000 (0:00:00.025) 0:04:46.300 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:27:13 +0000 (0:00:00.015) 0:04:46.316 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:27:13 +0000 (0:00:00.023) 0:04:46.339 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:27:16 +0000 (0:00:02.556) 0:04:48.896 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:27:18 +0000 (0:00:02.429) 0:04:51.325 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:27:18 +0000 (0:00:00.025) 0:04:51.350 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:27:18 +0000 (0:00:00.013) 0:04:51.364 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda1", "name": "luks-0ff83410-148e-4f69-832c-c589e3174417", "password": "-", "state": "absent" }, { "backing_device": "/dev/mapper/foo-test1", "name": "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "password": "-", "state": "present" } ], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" } ], "packages": [ "xfsprogs", "cryptsetup", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": "aes-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:27:28 +0000 (0:00:09.422) 0:05:00.786 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:27:28 +0000 (0:00:00.016) 0:05:00.802 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:27:28 +0000 (0:00:00.013) 0:05:00.816 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "luks" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create format", "device": "/dev/sda", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/foo", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/sda1", "name": "luks-0ff83410-148e-4f69-832c-c589e3174417", "password": "-", "state": "absent" }, { "backing_device": "/dev/mapper/foo-test1", "name": "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" } ], "packages": [ "xfsprogs", "cryptsetup", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": "aes-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:27:28 +0000 (0:00:00.019) 0:05:00.835 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": "aes-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:27:28 +0000 (0:00:00.018) 0:05:00.854 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:27:28 +0000 (0:00:00.016) 0:05:00.870 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-0ff83410-148e-4f69-832c-c589e3174417" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:27:28 +0000 (0:00:00.227) 0:05:01.098 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:27:29 +0000 (0:00:01.005) 0:05:02.103 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:27:29 +0000 (0:00:00.251) 0:05:02.354 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:27:29 +0000 (0:00:00.022) 0:05:02.377 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:27:30 +0000 (0:00:01.024) 0:05:03.401 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699209.3451245, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "1ef86e065a6271a3f2fe527f3c5c2df0a0eecfb5", "ctime": 1700699209.3431246, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699209.3421247, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 85, "uid": 0, "version": "4105456450", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:27:30 +0000 (0:00:00.226) 0:05:03.627 ***** changed: [sut] => (item={'backing_device': '/dev/sda1', 'name': 'luks-0ff83410-148e-4f69-832c-c589e3174417', 'password': '-', 'state': 'absent'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/sda1", "name": "luks-0ff83410-148e-4f69-832c-c589e3174417", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed changed: [sut] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'password': '-', 'state': 'present'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/mapper/foo-test1", "name": "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "password": "-", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:27:31 +0000 (0:00:00.458) 0:05:04.086 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:376 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.819) 0:05:04.905 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.025) 0:05:04.931 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": true, "encryption_cipher": "aes-xts-plain64", "encryption_key": null, "encryption_key_size": 512, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.018) 0:05:04.950 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.016) 0:05:04.966 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "1a8649e3-d3d1-4f83-afd2-dada9d2c560d" }, "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "size": "4G", "type": "crypt", "uuid": "619dfb64-19f8-4e06-8a2e-b4efd6cffec3" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.217) 0:05:05.183 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003348", "end": "2023-11-23 00:27:32.605155", "rc": 0, "start": "2023-11-23 00:27:32.601807" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.213) 0:05:05.396 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003332", "end": "2023-11-23 00:27:32.820428", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:27:32.817096" } STDOUT: luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.214) 0:05:05.611 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.030) 0:05:05.642 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.013) 0:05:05.655 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.034) 0:05:05.689 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:27:32 +0000 (0:00:00.020) 0:05:05.709 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.266) 0:05:05.976 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.017) 0:05:05.994 ***** ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.018) 0:05:06.013 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.019) 0:05:06.033 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.020) 0:05:06.054 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.018) 0:05:06.072 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.015) 0:05:06.088 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.024) 0:05:06.113 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.030) 0:05:06.144 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.018) 0:05:06.162 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.014) 0:05:06.176 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.014) 0:05:06.190 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.204 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.218 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.015) 0:05:06.234 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.014) 0:05:06.248 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.262 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.014) 0:05:06.276 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.290 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.304 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.030) 0:05:06.335 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:8 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.029) 0:05:06.365 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:16 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.014) 0:05:06.379 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:21 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.393 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:29 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.407 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:34 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.421 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:40 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.434 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:46 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.448 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.015) 0:05:06.463 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.030) 0:05:06.493 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:8 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.053) 0:05:06.547 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:16 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.014) 0:05:06.561 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:23 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.575 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:27 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.589 ***** ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.603 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.032) 0:05:06.635 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.016) 0:05:06.651 ***** skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.017) 0:05:06.669 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:2 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.027) 0:05:06.696 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:9 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.018) 0:05:06.715 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:18 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.017) 0:05:06.732 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:27 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.746 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:37 Thursday 23 November 2023 00:27:33 +0000 (0:00:00.013) 0:05:06.759 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:47 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.013) 0:05:06.773 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.013) 0:05:06.787 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.015) 0:05:06.803 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.034) 0:05:06.837 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:9 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.032) 0:05:06.870 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:16 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:06.884 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:22 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:06.898 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:28 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.013) 0:05:06.912 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:35 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.013) 0:05:06.926 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:41 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.015) 0:05:06.941 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:47 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:06.956 ***** ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:06.970 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.012) 0:05:06.983 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.026) 0:05:07.009 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.019) 0:05:07.029 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.073) 0:05:07.103 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.018) 0:05:07.121 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1014697, "block_size": 4096, "block_total": 1030144, "block_used": 15447, "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fstype": "xfs", "inode_available": 2093053, "inode_total": 2093056, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4156198912, "size_total": 4219469824, "uuid": "619dfb64-19f8-4e06-8a2e-b4efd6cffec3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1014697, "block_size": 4096, "block_total": 1030144, "block_used": 15447, "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fstype": "xfs", "inode_available": 2093053, "inode_total": 2093056, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4156198912, "size_total": 4219469824, "uuid": "619dfb64-19f8-4e06-8a2e-b4efd6cffec3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.024) 0:05:07.145 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.159 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.018) 0:05:07.178 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.018) 0:05:07.196 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.211 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.225 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.015) 0:05:07.241 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.018) 0:05:07.260 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.274 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.288 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.302 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.014) 0:05:07.317 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.032) 0:05:07.350 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.019) 0:05:07.369 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.017) 0:05:07.387 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.015) 0:05:07.402 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.013) 0:05:07.416 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.019) 0:05:07.435 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.052) 0:05:07.488 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699251.2975163, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699247.6656556, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1724, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699247.6656556, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.220) 0:05:07.708 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.019) 0:05:07.728 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.015) 0:05:07.743 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:27:34 +0000 (0:00:00.017) 0:05:07.760 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:27:35 +0000 (0:00:00.015) 0:05:07.776 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:27:35 +0000 (0:00:00.014) 0:05:07.791 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:27:35 +0000 (0:00:00.017) 0:05:07.808 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699251.281517, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699247.8916469, "dev": 5, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1760, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699247.8916469, "nlink": 1, "path": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:27:35 +0000 (0:00:00.220) 0:05:08.029 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:27:37 +0000 (0:00:02.510) 0:05:10.539 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/mapper/foo-test1" ], "delta": "0:00:00.007709", "end": "2023-11-23 00:27:37.966794", "rc": 0, "start": "2023-11-23 00:27:37.959085" } STDOUT: LUKS header information for /dev/mapper/foo-test1 Version: 1 Cipher name: aes Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 16384 MK bits: 512 MK digest: 7f 08 5e 9c 1f c8 ed 8e c9 72 3e 48 d8 85 62 ce 68 9e 8a f7 MK salt: 98 b3 3d e5 a4 f4 28 02 a2 f1 2e 2c 25 c4 5b 8c 2c f5 6e db fd b8 83 68 c2 0d 14 7b fb fb b6 64 MK iterations: 93223 UUID: 1a8649e3-d3d1-4f83-afd2-dada9d2c560d Key Slot 0: ENABLED Iterations: 1500108 Salt: ae 6a f6 65 c7 17 16 97 2d e5 f1 ce 3f 75 6e 70 43 2b 19 6f 4b 65 80 0d 58 19 04 74 8d 41 5f c4 Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:27:37 +0000 (0:00:00.217) 0:05:10.757 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.020) 0:05:10.778 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.021) 0:05:10.799 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.018) 0:05:10.817 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.017) 0:05:10.834 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.020) 0:05:10.855 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.024) 0:05:10.879 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.021) 0:05:10.901 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 -" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.018) 0:05:10.919 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.017) 0:05:10.937 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.018) 0:05:10.955 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.018) 0:05:10.974 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.024) 0:05:10.998 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.013 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.016) 0:05:11.029 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.043 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.015) 0:05:11.059 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.016) 0:05:11.075 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.016) 0:05:11.092 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.107 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.121 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.136 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.150 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.014) 0:05:11.165 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.305) 0:05:11.470 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.213) 0:05:11.683 ***** ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.020) 0:05:11.704 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:27:38 +0000 (0:00:00.017) 0:05:11.722 ***** ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.210) 0:05:11.932 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.018) 0:05:11.950 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.018) 0:05:11.969 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.018) 0:05:11.987 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.017) 0:05:12.004 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.018 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.033 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.013) 0:05:12.047 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.015) 0:05:12.062 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.076 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.013) 0:05:12.090 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.105 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.119 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.134 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.015) 0:05:12.149 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.013) 0:05:12.163 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.178 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.192 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.013) 0:05:12.206 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.220 ***** ok: [sut] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.016) 0:05:12.237 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.015) 0:05:12.252 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.023) 0:05:12.276 ***** ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.038110", "end": "2023-11-23 00:27:39.738251", "rc": 0, "start": "2023-11-23 00:27:39.700141" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.281) 0:05:12.557 ***** ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.020) 0:05:12.578 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.020) 0:05:12.598 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.015) 0:05:12.614 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.015) 0:05:12.629 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.015) 0:05:12.645 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.016) 0:05:12.662 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.014) 0:05:12.677 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.013) 0:05:12.690 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Verify preservation of encryption settings on existing LVM volume] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:379 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.013) 0:05:12.703 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.032) 0:05:12.736 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:27:39 +0000 (0:00:00.021) 0:05:12.757 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.017) 0:05:12.775 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.039) 0:05:12.814 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.014) 0:05:12.829 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.014) 0:05:12.843 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.013) 0:05:12.857 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.014) 0:05:12.872 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:27:40 +0000 (0:00:00.032) 0:05:12.904 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:27:42 +0000 (0:00:02.483) 0:05:15.387 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:27:42 +0000 (0:00:00.019) 0:05:15.407 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:27:42 +0000 (0:00:00.015) 0:05:15.423 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:27:44 +0000 (0:00:02.174) 0:05:17.597 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:27:44 +0000 (0:00:00.029) 0:05:17.626 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:27:44 +0000 (0:00:00.025) 0:05:17.651 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:27:44 +0000 (0:00:00.015) 0:05:17.666 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:27:44 +0000 (0:00:00.022) 0:05:17.689 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:27:47 +0000 (0:00:02.491) 0:05:20.181 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service": { "name": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:27:49 +0000 (0:00:02.470) 0:05:22.651 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:27:49 +0000 (0:00:00.026) 0:05:22.678 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d0ff83410\x2d148e\x2d4f69\x2d832c\x2dc589e3174417.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "name": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "systemd-udevd-kernel.socket systemd-journald.socket dev-sda1.device cryptsetup-pre.target -.mount \"system-systemd\\\\x2dcryptsetup.slice\" tmp.mount", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "\"blockdev@dev-mapper-luks\\\\x2d0ff83410\\\\x2d148e\\\\x2d4f69\\\\x2d832c\\\\x2dc589e3174417.target\" cryptsetup.target umount.target", "BindsTo": "dev-sda1.device", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-0ff83410-148e-4f69-832c-c589e3174417", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-0ff83410-148e-4f69-832c-c589e3174417 /dev/sda1 /tmp/storage_testw40hwjimlukskey ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-0ff83410-148e-4f69-832c-c589e3174417 /dev/sda1 /tmp/storage_testw40hwjimlukskey ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-0ff83410-148e-4f69-832c-c589e3174417 ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-0ff83410-148e-4f69-832c-c589e3174417 ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d0ff83410\\\\x2d148e\\\\x2d4f69\\\\x2d832c\\\\x2dc589e3174417.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target", "Requires": "tmp.mount \"system-systemd\\\\x2dcryptsetup.slice\" -.mount", "RequiresMountsFor": "/tmp/storage_testw40hwjimlukskey", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:27:30 UTC", "StateChangeTimestampMonotonic": "1185474896", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d0ff83410\\\\x2d148e\\\\x2d4f69\\\\x2d832c\\\\x2dc589e3174417.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:27:50 +0000 (0:00:01.050) 0:05:23.728 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2", "e2fsprogs", "cryptsetup" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:27:53 +0000 (0:00:02.230) 0:05:25.958 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:27:53 +0000 (0:00:00.015) 0:05:25.974 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d0ff83410\x2d148e\x2d4f69\x2d832c\x2dc589e3174417.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "name": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d0ff83410\\x2d148e\\x2d4f69\\x2d832c\\x2dc589e3174417.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d0ff83410\\\\x2d148e\\\\x2d4f69\\\\x2d832c\\\\x2dc589e3174417.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:27:54 +0000 (0:00:01.045) 0:05:27.019 ***** ok: [sut] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" ], "mounts": [ { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" } ], "packages": [ "xfsprogs", "lvm2", "e2fsprogs", "cryptsetup" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:27:54 +0000 (0:00:00.020) 0:05:27.040 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:27:54 +0000 (0:00:00.019) 0:05:27.060 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:27:54 +0000 (0:00:00.018) 0:05:27.079 ***** TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:27:54 +0000 (0:00:00.015) 0:05:27.094 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:27:55 +0000 (0:00:01.017) 0:05:28.112 ***** ok: [sut] => (item={'src': '/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:27:55 +0000 (0:00:00.229) 0:05:28.341 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:27:55 +0000 (0:00:00.021) 0:05:28.363 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:27:56 +0000 (0:00:01.018) 0:05:29.382 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699251.2965164, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "c5be5005f5e27506a9b1ec582e478cc0bc9faef0", "ctime": 1700699251.278517, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699251.277517, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 66, "uid": 0, "version": "3552488499", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:27:56 +0000 (0:00:00.221) 0:05:29.603 ***** TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:27:56 +0000 (0:00:00.014) 0:05:29.618 ***** ok: [sut] TASK [Assert preservation of encryption settings on existing LVM volume] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:393 Thursday 23 November 2023 00:27:57 +0000 (0:00:00.814) 0:05:30.432 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:400 Thursday 23 November 2023 00:27:57 +0000 (0:00:00.020) 0:05:30.453 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:27:57 +0000 (0:00:00.028) 0:05:30.481 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:27:57 +0000 (0:00:00.019) 0:05:30.500 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:27:57 +0000 (0:00:00.014) 0:05:30.515 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "1a8649e3-d3d1-4f83-afd2-dada9d2c560d" }, "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "size": "4G", "type": "crypt", "uuid": "619dfb64-19f8-4e06-8a2e-b4efd6cffec3" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:27:57 +0000 (0:00:00.217) 0:05:30.732 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003567", "end": "2023-11-23 00:27:58.154028", "rc": 0, "start": "2023-11-23 00:27:58.150461" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.211) 0:05:30.944 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003410", "end": "2023-11-23 00:27:58.365041", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:27:58.361631" } STDOUT: luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.210) 0:05:31.155 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.029) 0:05:31.185 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.199 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.031) 0:05:31.230 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.045) 0:05:31.276 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.212) 0:05:31.489 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.018) 0:05:31.507 ***** ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.019) 0:05:31.526 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.019) 0:05:31.546 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.017) 0:05:31.563 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.017) 0:05:31.581 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.013) 0:05:31.594 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.025) 0:05:31.620 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.028) 0:05:31.648 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.662 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.015) 0:05:31.678 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.693 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.707 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.722 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.736 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:27:58 +0000 (0:00:00.014) 0:05:31.750 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.015) 0:05:31.766 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.013) 0:05:31.780 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.013) 0:05:31.794 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:31.809 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.032) 0:05:31.841 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:8 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.028) 0:05:31.870 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:16 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:31.884 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:21 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.013) 0:05:31.898 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:29 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.015) 0:05:31.913 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:34 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.013) 0:05:31.927 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:40 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.015) 0:05:31.943 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:46 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:31.957 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:31.971 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.031) 0:05:32.002 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:8 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.029) 0:05:32.032 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:16 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.046 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:23 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.061 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:27 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.013) 0:05:32.075 ***** ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.013) 0:05:32.089 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.031) 0:05:32.120 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.018) 0:05:32.138 ***** skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.017) 0:05:32.156 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:2 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.027) 0:05:32.184 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:9 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.017) 0:05:32.202 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:18 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.017) 0:05:32.219 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:27 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.015) 0:05:32.235 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:37 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.249 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:47 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.264 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.278 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.292 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.059) 0:05:32.351 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:9 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.031) 0:05:32.382 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:16 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.015) 0:05:32.398 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:22 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.413 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:28 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.427 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:35 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.441 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:41 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.455 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:47 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.469 ***** ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.484 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.016) 0:05:32.500 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.029) 0:05:32.529 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.018) 0:05:32.547 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.075) 0:05:32.623 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.017) 0:05:32.640 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1014697, "block_size": 4096, "block_total": 1030144, "block_used": 15447, "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fstype": "xfs", "inode_available": 2093053, "inode_total": 2093056, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4156198912, "size_total": 4219469824, "uuid": "619dfb64-19f8-4e06-8a2e-b4efd6cffec3" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1014697, "block_size": 4096, "block_total": 1030144, "block_used": 15447, "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fstype": "xfs", "inode_available": 2093053, "inode_total": 2093056, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4156198912, "size_total": 4219469824, "uuid": "619dfb64-19f8-4e06-8a2e-b4efd6cffec3" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.023) 0:05:32.663 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.015) 0:05:32.679 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.018) 0:05:32.697 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.018) 0:05:32.715 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.730 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.744 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:27:59 +0000 (0:00:00.014) 0:05:32.759 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.018) 0:05:32.778 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.014) 0:05:32.792 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.015) 0:05:32.808 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.014) 0:05:32.823 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.014) 0:05:32.837 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.030) 0:05:32.867 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.018) 0:05:32.886 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.017) 0:05:32.903 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.015) 0:05:32.919 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.014) 0:05:32.934 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.019) 0:05:32.954 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.019) 0:05:32.973 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699257.9652607, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699247.6656556, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1724, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699247.6656556, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.219) 0:05:33.193 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.022) 0:05:33.215 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.016) 0:05:33.231 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.019) 0:05:33.251 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.022) 0:05:33.273 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.015) 0:05:33.288 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.021) 0:05:33.310 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699273.1136801, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699247.8916469, "dev": 5, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1760, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699247.8916469, "nlink": 1, "path": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:28:00 +0000 (0:00:00.225) 0:05:33.535 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:28:03 +0000 (0:00:02.560) 0:05:36.096 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/mapper/foo-test1" ], "delta": "0:00:00.007648", "end": "2023-11-23 00:28:03.526539", "rc": 0, "start": "2023-11-23 00:28:03.518891" } STDOUT: LUKS header information for /dev/mapper/foo-test1 Version: 1 Cipher name: aes Cipher mode: xts-plain64 Hash spec: sha256 Payload offset: 16384 MK bits: 512 MK digest: 7f 08 5e 9c 1f c8 ed 8e c9 72 3e 48 d8 85 62 ce 68 9e 8a f7 MK salt: 98 b3 3d e5 a4 f4 28 02 a2 f1 2e 2c 25 c4 5b 8c 2c f5 6e db fd b8 83 68 c2 0d 14 7b fb fb b6 64 MK iterations: 93223 UUID: 1a8649e3-d3d1-4f83-afd2-dada9d2c560d Key Slot 0: ENABLED Iterations: 1500108 Salt: ae 6a f6 65 c7 17 16 97 2d e5 f1 ce 3f 75 6e 70 43 2b 19 6f 4b 65 80 0d 58 19 04 74 8d 41 5f c4 Key material offset: 8 AF stripes: 4000 Key Slot 1: DISABLED Key Slot 2: DISABLED Key Slot 3: DISABLED Key Slot 4: DISABLED Key Slot 5: DISABLED Key Slot 6: DISABLED Key Slot 7: DISABLED TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.221) 0:05:36.317 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.020) 0:05:36.338 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.022) 0:05:36.361 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.018) 0:05:36.380 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.019) 0:05:36.399 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.022) 0:05:36.421 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.015) 0:05:36.437 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.015) 0:05:36.452 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 -" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.022) 0:05:36.475 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.018) 0:05:36.493 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.020) 0:05:36.513 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.019) 0:05:36.532 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.019) 0:05:36.551 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.014) 0:05:36.566 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.016) 0:05:36.582 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.014) 0:05:36.596 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.013) 0:05:36.610 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.013) 0:05:36.624 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.014) 0:05:36.638 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.014) 0:05:36.652 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.016) 0:05:36.669 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.014) 0:05:36.683 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.014) 0:05:36.697 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:28:03 +0000 (0:00:00.017) 0:05:36.714 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.214) 0:05:36.929 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.215) 0:05:37.145 ***** ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.020) 0:05:37.166 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.015) 0:05:37.181 ***** ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.216) 0:05:37.398 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.018) 0:05:37.416 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.017) 0:05:37.434 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.016) 0:05:37.450 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.017) 0:05:37.468 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.013) 0:05:37.482 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.015) 0:05:37.497 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.013) 0:05:37.511 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.014) 0:05:37.525 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.014) 0:05:37.540 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.014) 0:05:37.554 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.014) 0:05:37.569 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.015) 0:05:37.584 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.013) 0:05:37.598 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.013) 0:05:37.612 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.013) 0:05:37.626 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.013) 0:05:37.640 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.014) 0:05:37.654 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.016) 0:05:37.670 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.014) 0:05:37.685 ***** ok: [sut] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.016) 0:05:37.701 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.015) 0:05:37.717 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:28:04 +0000 (0:00:00.020) 0:05:37.738 ***** ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.034212", "end": "2023-11-23 00:28:05.198432", "rc": 0, "start": "2023-11-23 00:28:05.164220" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.251) 0:05:37.990 ***** ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.020) 0:05:38.011 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.020) 0:05:38.031 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.018) 0:05:38.049 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.018) 0:05:38.067 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.021) 0:05:38.089 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.017) 0:05:38.106 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.014) 0:05:38.120 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.017) 0:05:38.138 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml:12 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.016) 0:05:38.155 ***** changed: [sut] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Test for correct handling of safe_mode] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:406 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.225) 0:05:38.380 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.030) 0:05:38.411 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.020) 0:05:38.431 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.022) 0:05:38.453 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.051) 0:05:38.505 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.018) 0:05:38.523 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.037) 0:05:38.561 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.014) 0:05:38.576 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.015) 0:05:38.591 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.014) 0:05:38.606 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.014) 0:05:38.620 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:28:05 +0000 (0:00:00.034) 0:05:38.655 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:28:08 +0000 (0:00:02.535) 0:05:41.190 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "encryption": false, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:28:08 +0000 (0:00:00.021) 0:05:41.212 ***** ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:28:08 +0000 (0:00:00.021) 0:05:41.234 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:28:10 +0000 (0:00:02.182) 0:05:43.416 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:28:10 +0000 (0:00:00.028) 0:05:43.445 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:28:10 +0000 (0:00:00.023) 0:05:43.469 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:28:10 +0000 (0:00:00.014) 0:05:43.484 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:28:10 +0000 (0:00:00.023) 0:05:43.508 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:28:13 +0000 (0:00:02.495) 0:05:46.004 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service": { "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:28:15 +0000 (0:00:02.456) 0:05:48.460 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:28:15 +0000 (0:00:00.025) 0:05:48.485 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d1a8649e3\x2dd3d1\x2d4f83\x2dafd2\x2ddada9d2c560d.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "cryptsetup-pre.target systemd-udevd-kernel.socket \"system-systemd\\\\x2dcryptsetup.slice\" \"dev-mapper-foo\\\\x2dtest1.device\" systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "cryptsetup.target umount.target \"blockdev@dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.target\"", "BindsTo": "\"dev-mapper-foo\\\\x2dtest1.device\"", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.service\"", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target \"dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.device\"", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:27:54 UTC", "StateChangeTimestampMonotonic": "1209106209", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:28:16 +0000 (0:00:01.063) 0:05:49.548 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting on device 'luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d' in safe mode due to encryption removal TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:28:18 +0000 (0:00:02.182) 0:05:51.731 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting on device 'luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d' in safe mode due to encryption removal", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': 0, 'encryption_luks_version': 'luks1', 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:28:18 +0000 (0:00:00.020) 0:05:51.751 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d1a8649e3\x2dd3d1\x2d4f83\x2dafd2\x2ddada9d2c560d.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target \"dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.device\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:27:54 UTC", "StateChangeTimestampMonotonic": "1209106209", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:28:20 +0000 (0:00:01.037) 0:05:52.789 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.048) 0:05:52.837 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.023) 0:05:52.861 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:11 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.014) 0:05:52.875 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699285.5902019, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699285.5902019, "dev": 64769, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1700699285.5902019, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "799404816", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Assert file presence] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:16 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.223) 0:05:53.098 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Remove the encryption layer] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:429 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.018) 0:05:53.116 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.041) 0:05:53.158 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.021) 0:05:53.180 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.017) 0:05:53.198 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.039) 0:05:53.237 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.014) 0:05:53.252 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.014) 0:05:53.266 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.014) 0:05:53.281 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.014) 0:05:53.295 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:28:20 +0000 (0:00:00.033) 0:05:53.329 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:28:23 +0000 (0:00:02.517) 0:05:55.847 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "encryption": false, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:28:23 +0000 (0:00:00.018) 0:05:55.865 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:28:23 +0000 (0:00:00.015) 0:05:55.880 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:28:25 +0000 (0:00:02.145) 0:05:58.025 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:28:25 +0000 (0:00:00.027) 0:05:58.053 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:28:25 +0000 (0:00:00.024) 0:05:58.078 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:28:25 +0000 (0:00:00.016) 0:05:58.094 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:28:25 +0000 (0:00:00.027) 0:05:58.122 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:28:27 +0000 (0:00:02.473) 0:06:00.595 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service": { "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:28:30 +0000 (0:00:02.408) 0:06:03.004 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:28:30 +0000 (0:00:00.025) 0:06:03.030 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d1a8649e3\x2dd3d1\x2d4f83\x2dafd2\x2ddada9d2c560d.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "\"dev-mapper-foo\\\\x2dtest1.device\" cryptsetup-pre.target systemd-udevd-kernel.socket systemd-journald.socket \"system-systemd\\\\x2dcryptsetup.slice\"", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "\"blockdev@dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.target\" umount.target cryptsetup.target", "BindsTo": "\"dev-mapper-foo\\\\x2dtest1.device\"", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.service\"", "NeedDaemonReload": "no", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "\"dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.device\" cryptsetup.target", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:27:54 UTC", "StateChangeTimestampMonotonic": "1209106209", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:28:31 +0000 (0:00:01.034) 0:06:04.064 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/mapper/foo-test1", "name": "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "password": "-", "state": "absent" } ], "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:28:34 +0000 (0:00:03.071) 0:06:07.135 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:28:34 +0000 (0:00:00.017) 0:06:07.152 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d1a8649e3\x2dd3d1\x2d4f83\x2dafd2\x2ddada9d2c560d.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "RequiredBy": "\"dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.device\" cryptsetup.target", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:27:54 UTC", "StateChangeTimestampMonotonic": "1209106209", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:28:35 +0000 (0:00:01.034) 0:06:08.187 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/mapper/foo-test1", "name": "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/mapper/foo-test1", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" } ], "packages": [ "e2fsprogs", "lvm2", "xfsprogs" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:28:35 +0000 (0:00:00.022) 0:06:08.210 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:28:35 +0000 (0:00:00.021) 0:06:08.231 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:28:35 +0000 (0:00:00.019) 0:06:08.250 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:28:35 +0000 (0:00:00.236) 0:06:08.487 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:28:36 +0000 (0:00:01.004) 0:06:09.492 ***** changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:28:36 +0000 (0:00:00.244) 0:06:09.736 ***** skipping: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:28:36 +0000 (0:00:00.020) 0:06:09.757 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:28:37 +0000 (0:00:00.996) 0:06:10.753 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699251.2965164, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "c5be5005f5e27506a9b1ec582e478cc0bc9faef0", "ctime": 1700699251.278517, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699251.277517, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 66, "uid": 0, "version": "3552488499", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:28:38 +0000 (0:00:00.220) 0:06:10.974 ***** changed: [sut] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d', 'password': '-', 'state': 'absent'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/mapper/foo-test1", "name": "luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:28:38 +0000 (0:00:00.238) 0:06:11.213 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:445 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.813) 0:06:12.026 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.031) 0:06:12.058 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/foo-test1", "_kernel_device": "/dev/dm-0", "_mount_id": "/dev/mapper/foo-test1", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": 0, "encryption_luks_version": "luks1", "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.020) 0:06:12.079 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.014) 0:06:12.093 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "76602fd5-e92d-4133-bf92-c34d474e6aad" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.214) 0:06:12.308 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003558", "end": "2023-11-23 00:28:39.732097", "rc": 0, "start": "2023-11-23 00:28:39.728539" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/foo-test1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.215) 0:06:12.523 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003371", "end": "2023-11-23 00:28:39.946737", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:28:39.943366" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:28:39 +0000 (0:00:00.213) 0:06:12.737 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.030) 0:06:12.767 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.013) 0:06:12.781 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.030) 0:06:12.811 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.047) 0:06:12.859 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.213) 0:06:13.073 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.018) 0:06:13.091 ***** ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.018) 0:06:13.110 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.019) 0:06:13.130 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.017) 0:06:13.148 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.017) 0:06:13.165 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.180 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.022) 0:06:13.202 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.030) 0:06:13.232 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.246 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.013) 0:06:13.260 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.013) 0:06:13.274 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.289 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.013) 0:06:13.303 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.318 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.332 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.346 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.360 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.015) 0:06:13.376 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.390 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.028) 0:06:13.419 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:8 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.032) 0:06:13.451 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:16 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.015) 0:06:13.466 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:21 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.015) 0:06:13.482 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:29 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.497 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:34 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.512 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:40 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.016) 0:06:13.528 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:46 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.542 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.014) 0:06:13.557 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.031) 0:06:13.589 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:8 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.032) 0:06:13.621 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:16 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.019) 0:06:13.640 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:23 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.015) 0:06:13.656 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:27 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.013) 0:06:13.670 ***** ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.018) 0:06:13.688 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.036) 0:06:13.725 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.017) 0:06:13.743 ***** skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:28:40 +0000 (0:00:00.017) 0:06:13.760 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:2 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.029) 0:06:13.790 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:9 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.023) 0:06:13.813 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:18 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.020) 0:06:13.833 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:27 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:13.849 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:37 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:13.865 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:47 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.017) 0:06:13.883 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:13.898 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:13.912 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.033) 0:06:13.946 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:9 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.060) 0:06:14.007 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:16 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:14.023 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:22 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:14.038 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:28 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.053 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:35 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.067 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:41 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:14.082 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:47 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:14.097 ***** ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.013) 0:06:14.111 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.013) 0:06:14.125 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.029) 0:06:14.154 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.017) 0:06:14.172 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.073) 0:06:14.246 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/foo-test1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.018) 0:06:14.264 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1016730, "block_size": 4096, "block_total": 1032192, "block_used": 15462, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4164526080, "size_total": 4227858432, "uuid": "76602fd5-e92d-4133-bf92-c34d474e6aad" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1016730, "block_size": 4096, "block_total": 1032192, "block_used": 15462, "device": "/dev/mapper/foo-test1", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4164526080, "size_total": 4227858432, "uuid": "76602fd5-e92d-4133-bf92-c34d474e6aad" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.026) 0:06:14.290 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:14.306 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.018) 0:06:14.324 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.017) 0:06:14.342 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.356 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.015) 0:06:14.372 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.386 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.018) 0:06:14.405 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.419 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.434 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.448 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.016) 0:06:14.464 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/foo-test1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.031) 0:06:14.496 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.018) 0:06:14.514 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.017) 0:06:14.532 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.546 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.014) 0:06:14.561 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.023) 0:06:14.585 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:28:41 +0000 (0:00:00.021) 0:06:14.606 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699318.4099436, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699314.2621026, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1899, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699314.2621026, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.220) 0:06:14.827 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.019) 0:06:14.846 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.014) 0:06:14.861 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.017) 0:06:14.878 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.015) 0:06:14.893 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.014) 0:06:14.908 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.018) 0:06:14.926 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:28:42 +0000 (0:00:00.014) 0:06:14.940 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:28:44 +0000 (0:00:02.480) 0:06:17.421 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.042) 0:06:17.464 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.015) 0:06:17.479 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.020) 0:06:17.500 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.514 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.529 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.013) 0:06:17.543 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.015) 0:06:17.558 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.573 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.018) 0:06:17.592 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.017) 0:06:17.610 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.624 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.638 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.016) 0:06:17.655 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.015) 0:06:17.670 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.017) 0:06:17.688 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.702 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.015) 0:06:17.717 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.013) 0:06:17.731 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.017) 0:06:17.748 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:28:44 +0000 (0:00:00.014) 0:06:17.762 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.015) 0:06:17.778 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.016) 0:06:17.794 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:17.809 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:17.823 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.209) 0:06:18.032 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.212) 0:06:18.244 ***** ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.020) 0:06:18.265 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.016) 0:06:18.281 ***** ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.209) 0:06:18.491 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.018) 0:06:18.509 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.017) 0:06:18.527 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.017) 0:06:18.545 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.018) 0:06:18.563 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.016) 0:06:18.579 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:18.594 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:18.608 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:18.622 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:18.636 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.016) 0:06:18.653 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.017) 0:06:18.670 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.015) 0:06:18.686 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.015) 0:06:18.701 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.016) 0:06:18.717 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.015) 0:06:18.733 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.014) 0:06:18.747 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:28:45 +0000 (0:00:00.015) 0:06:18.763 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.014) 0:06:18.777 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.015) 0:06:18.793 ***** ok: [sut] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.017) 0:06:18.810 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.015) 0:06:18.825 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.021) 0:06:18.847 ***** ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.042369", "end": "2023-11-23 00:28:46.311100", "rc": 0, "start": "2023-11-23 00:28:46.268731" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.257) 0:06:19.105 ***** ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.019) 0:06:19.124 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.021) 0:06:19.146 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.015) 0:06:19.161 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.016) 0:06:19.178 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.015) 0:06:19.194 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.016) 0:06:19.210 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.013) 0:06:19.224 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.014) 0:06:19.238 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create a file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/create-test-file.yml:12 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.014) 0:06:19.252 ***** changed: [sut] => { "changed": true, "dest": "/opt/test1/quux", "gid": 0, "group": "root", "mode": "0644", "owner": "root", "secontext": "unconfined_u:object_r:unlabeled_t:s0", "size": 0, "state": "file", "uid": 0 } TASK [Test for correct handling of safe_mode] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:451 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.219) 0:06:19.472 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml for sut TASK [Store global variable value copy] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:4 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.030) 0:06:19.503 ***** ok: [sut] => { "ansible_facts": { "storage_pools_global": [], "storage_safe_mode_global": true, "storage_volumes_global": [] }, "changed": false } TASK [Verify role raises correct error] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:10 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.017) 0:06:19.520 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.056) 0:06:19.576 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.022) 0:06:19.599 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.019) 0:06:19.619 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.037) 0:06:19.656 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.015) 0:06:19.671 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.014) 0:06:19.686 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.014) 0:06:19.700 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.013) 0:06:19.714 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:28:46 +0000 (0:00:00.032) 0:06:19.746 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:28:49 +0000 (0:00:02.504) 0:06:22.251 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:28:49 +0000 (0:00:00.020) 0:06:22.272 ***** ok: [sut] => { "storage_volumes": [] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:28:49 +0000 (0:00:00.018) 0:06:22.290 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup", "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:28:51 +0000 (0:00:02.007) 0:06:24.298 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:28:51 +0000 (0:00:00.027) 0:06:24.325 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:28:51 +0000 (0:00:00.023) 0:06:24.349 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:28:51 +0000 (0:00:00.014) 0:06:24.363 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:28:51 +0000 (0:00:00.021) 0:06:24.385 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:28:54 +0000 (0:00:02.476) 0:06:26.861 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service": { "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "source": "systemd", "state": "stopped", "status": "generated" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:28:56 +0000 (0:00:02.442) 0:06:29.304 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [ "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service" ] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:28:56 +0000 (0:00:00.023) 0:06:29.327 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d1a8649e3\x2dd3d1\x2d4f83\x2dafd2\x2ddada9d2c560d.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "status": { "AccessSELinuxContext": "system_u:object_r:systemd_unit_file_t:s0", "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "After": "\"dev-mapper-foo\\\\x2dtest1.device\" \"system-systemd\\\\x2dcryptsetup.slice\" systemd-udevd-kernel.socket cryptsetup-pre.target systemd-journald.socket", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "Before": "cryptsetup.target umount.target \"blockdev@dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.target\"", "BindsTo": "\"dev-mapper-foo\\\\x2dtest1.device\"", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "yes", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "Conflicts": "umount.target", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "no", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "Cryptography Setup for luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d", "DevicePolicy": "auto", "Documentation": "\"man:crypttab(5)\" \"man:systemd-cryptsetup-generator(8)\" \"man:systemd-cryptsetup@.service(8)\"", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExecStart": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStartEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup attach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d /dev/mapper/foo-test1 - ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStop": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d ; ignore_errors=no ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExecStopEx": "{ path=/usr/lib/systemd/systemd-cryptsetup ; argv[]=/usr/lib/systemd/systemd-cryptsetup detach luks-1a8649e3-d3d1-4f83-afd2-dada9d2c560d ; flags= ; start_time=[n/a] ; stop_time=[n/a] ; pid=0 ; code=(null) ; status=0/0 }", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/run/systemd/generator/systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "IgnoreOnIsolate": "yes", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "shared", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "8388608", "LimitMEMLOCKSoft": "8388608", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "524288", "LimitNOFILESoft": "1024", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadState": "loaded", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMPolicy": "stop", "OOMScoreAdjust": "500", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "yes", "RemoveIPC": "no", "RequiredBy": "cryptsetup.target", "Requires": "\"system-systemd\\\\x2dcryptsetup.slice\"", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "Slice": "system-systemd\\x2dcryptsetup.slice", "SourcePath": "/etc/crypttab", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "journal", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestamp": "Thu 2023-11-23 00:27:54 UTC", "StateChangeTimestampMonotonic": "1209106209", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "infinity", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "infinity", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "infinity", "TimerSlackNSec": "50000", "Transient": "no", "Type": "oneshot", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "generated", "UtmpMode": "init", "Wants": "\"blockdev@dev-mapper-luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.target\"", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:28:57 +0000 (0:00:01.028) 0:06:30.356 ***** fatal: [sut]: FAILED! => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } MSG: cannot remove existing formatting on device 'test1' in safe mode due to adding encryption TASK [linux-system-roles.storage : Failed message] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:106 Thursday 23 November 2023 00:28:59 +0000 (0:00:02.101) 0:06:32.458 ***** fatal: [sut]: FAILED! => { "changed": false } MSG: {'msg': "cannot remove existing formatting on device 'test1' in safe mode due to adding encryption", 'changed': False, 'actions': [], 'leaves': [], 'mounts': [], 'crypts': [], 'pools': [], 'volumes': [], 'packages': [], 'failed': True, 'invocation': {'module_args': {'pools': [{'disks': ['sda'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'foo', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'raid_chunk_size': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': True, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': 'yabbadabbadoo', 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'mount_user': None, 'mount_group': None, 'mount_mode': None, 'name': 'test1', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'raid_stripe_size': None, 'thin_pool_name': None, 'thin_pool_size': None, 'thin': False, 'vdo_pool_size': None, 'disks': ['sda'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}]}], 'volumes': [], 'use_partitions': None, 'disklabel_type': None, 'pool_defaults': {'state': 'present', 'type': 'lvm', 'disks': [], 'volumes': [], 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None}, 'volume_defaults': {'state': 'present', 'type': 'lvm', 'size': 0, 'disks': [], 'fs_type': 'xfs', 'fs_label': '', 'fs_create_options': '', 'fs_overwrite_existing': True, 'mount_point': '', 'mount_options': 'defaults', 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_stripe_size': None, 'raid_metadata_version': None, 'encryption': False, 'encryption_password': None, 'encryption_key': None, 'encryption_cipher': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'compression': None, 'deduplication': None, 'vdo_pool_size': None, 'thin': None, 'thin_pool_name': None, 'thin_pool_size': None, 'cached': False, 'cache_size': 0, 'cache_mode': None, 'cache_devices': []}, 'safe_mode': True, 'packages_only': False, 'diskvolume_mkfs_option_map': {}}}, '_ansible_no_log': False} TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:28:59 +0000 (0:00:00.019) 0:06:32.477 ***** changed: [sut] => (item=systemd-cryptsetup@luks\x2d1a8649e3\x2dd3d1\x2d4f83\x2dafd2\x2ddada9d2c560d.service) => { "ansible_loop_var": "item", "changed": true, "item": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "name": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "status": { "ActiveEnterTimestampMonotonic": "0", "ActiveExitTimestampMonotonic": "0", "ActiveState": "inactive", "AllowIsolate": "no", "AssertResult": "no", "AssertTimestampMonotonic": "0", "BlockIOAccounting": "no", "BlockIOWeight": "[not set]", "CPUAccounting": "yes", "CPUAffinityFromNUMA": "no", "CPUQuotaPerSecUSec": "infinity", "CPUQuotaPeriodUSec": "infinity", "CPUSchedulingPolicy": "0", "CPUSchedulingPriority": "0", "CPUSchedulingResetOnFork": "no", "CPUShares": "[not set]", "CPUUsageNSec": "[not set]", "CPUWeight": "[not set]", "CacheDirectoryMode": "0755", "CanFreeze": "yes", "CanIsolate": "no", "CanReload": "no", "CanStart": "no", "CanStop": "yes", "CapabilityBoundingSet": "cap_chown cap_dac_override cap_dac_read_search cap_fowner cap_fsetid cap_kill cap_setgid cap_setuid cap_setpcap cap_linux_immutable cap_net_bind_service cap_net_broadcast cap_net_admin cap_net_raw cap_ipc_lock cap_ipc_owner cap_sys_module cap_sys_rawio cap_sys_chroot cap_sys_ptrace cap_sys_pacct cap_sys_admin cap_sys_boot cap_sys_nice cap_sys_resource cap_sys_time cap_sys_tty_config cap_mknod cap_lease cap_audit_write cap_audit_control cap_setfcap cap_mac_override cap_mac_admin cap_syslog cap_wake_alarm cap_block_suspend cap_audit_read cap_perfmon cap_bpf cap_checkpoint_restore", "CleanResult": "success", "CollectMode": "inactive", "ConditionResult": "no", "ConditionTimestampMonotonic": "0", "ConfigurationDirectoryMode": "0755", "ControlGroupId": "0", "ControlPID": "0", "CoredumpFilter": "0x33", "DefaultDependencies": "yes", "DefaultMemoryLow": "0", "DefaultMemoryMin": "0", "Delegate": "no", "Description": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "DevicePolicy": "auto", "DropInPaths": "/usr/lib/systemd/system/service.d/10-timeout-abort.conf", "DynamicUser": "no", "ExecMainCode": "0", "ExecMainExitTimestampMonotonic": "0", "ExecMainPID": "0", "ExecMainStartTimestampMonotonic": "0", "ExecMainStatus": "0", "ExitType": "main", "FailureAction": "none", "FileDescriptorStoreMax": "0", "FinalKillSignal": "9", "FragmentPath": "/etc/systemd/system/systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "FreezerState": "running", "GID": "[not set]", "GuessMainPID": "yes", "IOAccounting": "no", "IOReadBytes": "18446744073709551615", "IOReadOperations": "18446744073709551615", "IOSchedulingClass": "2", "IOSchedulingPriority": "4", "IOWeight": "[not set]", "IOWriteBytes": "18446744073709551615", "IOWriteOperations": "18446744073709551615", "IPAccounting": "no", "IPEgressBytes": "[no data]", "IPEgressPackets": "[no data]", "IPIngressBytes": "[no data]", "IPIngressPackets": "[no data]", "Id": "systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service", "IgnoreOnIsolate": "no", "IgnoreSIGPIPE": "yes", "InactiveEnterTimestampMonotonic": "0", "InactiveExitTimestampMonotonic": "0", "JobRunningTimeoutUSec": "infinity", "JobTimeoutAction": "none", "JobTimeoutUSec": "infinity", "KeyringMode": "private", "KillMode": "control-group", "KillSignal": "15", "LimitAS": "infinity", "LimitASSoft": "infinity", "LimitCORE": "infinity", "LimitCORESoft": "infinity", "LimitCPU": "infinity", "LimitCPUSoft": "infinity", "LimitDATA": "infinity", "LimitDATASoft": "infinity", "LimitFSIZE": "infinity", "LimitFSIZESoft": "infinity", "LimitLOCKS": "infinity", "LimitLOCKSSoft": "infinity", "LimitMEMLOCK": "486424576", "LimitMEMLOCKSoft": "486424576", "LimitMSGQUEUE": "819200", "LimitMSGQUEUESoft": "819200", "LimitNICE": "0", "LimitNICESoft": "0", "LimitNOFILE": "1073741816", "LimitNOFILESoft": "1073741816", "LimitNPROC": "14734", "LimitNPROCSoft": "14734", "LimitRSS": "infinity", "LimitRSSSoft": "infinity", "LimitRTPRIO": "0", "LimitRTPRIOSoft": "0", "LimitRTTIME": "infinity", "LimitRTTIMESoft": "infinity", "LimitSIGPENDING": "14734", "LimitSIGPENDINGSoft": "14734", "LimitSTACK": "infinity", "LimitSTACKSoft": "8388608", "LoadError": "org.freedesktop.systemd1.UnitMasked \"Unit systemd-cryptsetup@luks\\x2d1a8649e3\\x2dd3d1\\x2d4f83\\x2dafd2\\x2ddada9d2c560d.service is masked.\"", "LoadState": "masked", "LockPersonality": "no", "LogLevelMax": "-1", "LogRateLimitBurst": "0", "LogRateLimitIntervalUSec": "0", "LogsDirectoryMode": "0755", "MainPID": "0", "ManagedOOMMemoryPressure": "auto", "ManagedOOMMemoryPressureLimit": "0", "ManagedOOMPreference": "none", "ManagedOOMSwap": "auto", "MemoryAccounting": "yes", "MemoryAvailable": "infinity", "MemoryCurrent": "[not set]", "MemoryDenyWriteExecute": "no", "MemoryHigh": "infinity", "MemoryLimit": "infinity", "MemoryLow": "0", "MemoryMax": "infinity", "MemoryMin": "0", "MemorySwapMax": "infinity", "MemoryZSwapMax": "infinity", "MountAPIVFS": "no", "NFileDescriptorStore": "0", "NRestarts": "0", "NUMAPolicy": "n/a", "Names": "\"systemd-cryptsetup@luks\\\\x2d1a8649e3\\\\x2dd3d1\\\\x2d4f83\\\\x2dafd2\\\\x2ddada9d2c560d.service\"", "NeedDaemonReload": "yes", "Nice": "0", "NoNewPrivileges": "no", "NonBlocking": "no", "NotifyAccess": "none", "OOMScoreAdjust": "0", "OnFailureJobMode": "replace", "OnSuccessJobMode": "fail", "Perpetual": "no", "PrivateDevices": "no", "PrivateIPC": "no", "PrivateMounts": "no", "PrivateNetwork": "no", "PrivateTmp": "no", "PrivateUsers": "no", "ProcSubset": "all", "ProtectClock": "no", "ProtectControlGroups": "no", "ProtectHome": "no", "ProtectHostname": "no", "ProtectKernelLogs": "no", "ProtectKernelModules": "no", "ProtectKernelTunables": "no", "ProtectProc": "default", "ProtectSystem": "no", "RefuseManualStart": "no", "RefuseManualStop": "no", "ReloadResult": "success", "ReloadSignal": "1", "RemainAfterExit": "no", "RemoveIPC": "no", "Restart": "no", "RestartKillSignal": "15", "RestartUSec": "100ms", "RestrictNamespaces": "no", "RestrictRealtime": "no", "RestrictSUIDSGID": "no", "Result": "success", "RootDirectoryStartOnly": "no", "RuntimeDirectoryMode": "0755", "RuntimeDirectoryPreserve": "no", "RuntimeMaxUSec": "infinity", "RuntimeRandomizedExtraUSec": "0", "SameProcessGroup": "no", "SecureBits": "0", "SendSIGHUP": "no", "SendSIGKILL": "yes", "StandardError": "inherit", "StandardInput": "null", "StandardOutput": "inherit", "StartLimitAction": "none", "StartLimitBurst": "5", "StartLimitIntervalUSec": "10s", "StartupBlockIOWeight": "[not set]", "StartupCPUShares": "[not set]", "StartupCPUWeight": "[not set]", "StartupIOWeight": "[not set]", "StateChangeTimestampMonotonic": "0", "StateDirectoryMode": "0755", "StatusErrno": "0", "StopWhenUnneeded": "no", "SubState": "dead", "SuccessAction": "none", "SyslogFacility": "3", "SyslogLevel": "6", "SyslogLevelPrefix": "yes", "SyslogPriority": "30", "SystemCallErrorNumber": "2147483646", "TTYReset": "no", "TTYVHangup": "no", "TTYVTDisallocate": "no", "TasksAccounting": "yes", "TasksCurrent": "[not set]", "TasksMax": "4420", "TimeoutAbortUSec": "45s", "TimeoutCleanUSec": "infinity", "TimeoutStartFailureMode": "terminate", "TimeoutStartUSec": "45s", "TimeoutStopFailureMode": "abort", "TimeoutStopUSec": "45s", "TimerSlackNSec": "50000", "Transient": "no", "UID": "[not set]", "UMask": "0022", "UnitFilePreset": "disabled", "UnitFileState": "masked", "UtmpMode": "init", "WatchdogSignal": "6", "WatchdogTimestampMonotonic": "0", "WatchdogUSec": "infinity" } } TASK [Check that we failed in the role] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:29 Thursday 23 November 2023 00:29:00 +0000 (0:00:01.026) 0:06:33.504 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the blivet output and error message are correct] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:34 Thursday 23 November 2023 00:29:00 +0000 (0:00:00.017) 0:06:33.522 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify correct exception or error message] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-failed.yml:45 Thursday 23 November 2023 00:29:00 +0000 (0:00:00.022) 0:06:33.544 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the file] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:11 Thursday 23 November 2023 00:29:00 +0000 (0:00:00.014) 0:06:33.558 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699326.6806266, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699326.6806266, "dev": 64768, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 131, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0644", "mtime": 1700699326.6806266, "nlink": 1, "path": "/opt/test1/quux", "pw_name": "root", "readable": true, "rgrp": true, "roth": true, "rusr": true, "size": 0, "uid": 0, "version": "132750507", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Assert file presence] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-data-preservation.yml:16 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.250) 0:06:33.808 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Add encryption to the volume] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:474 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.018) 0:06:33.827 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.045) 0:06:33.872 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.021) 0:06:33.894 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.017) 0:06:33.912 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.037) 0:06:33.950 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.014) 0:06:33.965 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.014) 0:06:33.979 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.013) 0:06:33.992 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.013) 0:06:34.006 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:29:01 +0000 (0:00:00.030) 0:06:34.037 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:29:03 +0000 (0:00:02.521) 0:06:36.559 ***** ok: [sut] => { "storage_pools": [ { "disks": [ "sda" ], "name": "foo", "type": "lvm", "volumes": [ { "encryption": true, "encryption_password": "yabbadabbadoo", "mount_point": "/opt/test1", "name": "test1", "size": "4g" } ] } ] } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:29:03 +0000 (0:00:00.018) 0:06:36.577 ***** ok: [sut] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:29:03 +0000 (0:00:00.016) 0:06:36.593 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "cryptsetup", "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:29:05 +0000 (0:00:02.004) 0:06:38.598 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:29:05 +0000 (0:00:00.028) 0:06:38.627 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:29:05 +0000 (0:00:00.022) 0:06:38.650 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:29:05 +0000 (0:00:00.014) 0:06:38.664 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:29:05 +0000 (0:00:00.022) 0:06:38.687 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:29:08 +0000 (0:00:02.520) 0:06:41.208 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:29:10 +0000 (0:00:02.452) 0:06:43.660 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:29:10 +0000 (0:00:00.025) 0:06:43.685 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:29:10 +0000 (0:00:00.012) 0:06:43.698 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/mapper/foo-test1", "name": "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "password": "-", "state": "present" } ], "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs", "xfsprogs", "cryptsetup" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:29:22 +0000 (0:00:11.160) 0:06:54.859 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:29:22 +0000 (0:00:00.016) 0:06:54.875 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:29:22 +0000 (0:00:00.013) 0:06:54.889 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "xfs" }, { "action": "create format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "create device", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": "xfs" } ], "changed": true, "crypts": [ { "backing_device": "/dev/mapper/foo-test1", "name": "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "password": "-", "state": "present" } ], "failed": false, "leaves": [ "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0", "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "mounted" } ], "packages": [ "lvm2", "e2fsprogs", "xfsprogs", "cryptsetup" ], "pools": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:29:22 +0000 (0:00:00.019) 0:06:54.909 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:29:22 +0000 (0:00:00.017) 0:06:54.926 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:29:22 +0000 (0:00:00.017) 0:06:54.944 ***** changed: [sut] => (item={'src': '/dev/mapper/foo-test1', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/foo-test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/foo-test1" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:29:22 +0000 (0:00:00.229) 0:06:55.173 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:29:23 +0000 (0:00:01.031) 0:06:56.205 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164" } TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:29:23 +0000 (0:00:00.254) 0:06:56.460 ***** skipping: [sut] => (item={'src': '/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted', 'owner': None, 'group': None, 'mode': None}) => { "ansible_loop_var": "mount_info", "changed": false, "mount_info": { "dump": 0, "fstype": "xfs", "group": null, "mode": null, "opts": "defaults", "owner": null, "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "mounted" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:29:23 +0000 (0:00:00.020) 0:06:56.480 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:29:24 +0000 (0:00:01.014) 0:06:57.495 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699319.9448848, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 0, "charset": "binary", "checksum": "da39a3ee5e6b4b0d3255bfef95601890afd80709", "ctime": 1700699318.4039438, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263320, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "inode/x-empty", "mode": "0600", "mtime": 1700699318.4029438, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": "525494302", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:29:24 +0000 (0:00:00.218) 0:06:57.714 ***** changed: [sut] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164', 'password': '-', 'state': 'present'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/mapper/foo-test1", "name": "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "password": "-", "state": "present" } } MSG: line added TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:29:25 +0000 (0:00:00.245) 0:06:57.959 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:490 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.814) 0:06:58.774 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.035) 0:06:58.809 ***** ok: [sut] => { "_storage_pools_list": [ { "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_kernel_device": "/dev/dm-1", "_mount_id": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "_raw_device": "/dev/mapper/foo-test1", "_raw_kernel_device": "/dev/dm-0", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": true, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": "yabbadabbadoo", "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "mount_user": null, "name": "test1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": "4g", "state": "present", "thin": false, "thin_pool_name": null, "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.019) 0:06:58.829 ***** skipping: [sut] => {} TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.014) 0:06:58.843 ***** ok: [sut] => { "changed": false, "info": { "/dev/mapper/foo-test1": { "fstype": "crypto_LUKS", "label": "", "name": "/dev/mapper/foo-test1", "size": "4G", "type": "lvm", "uuid": "87ddbe6c-ff42-47dc-9b03-3e5b4e379164" }, "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164": { "fstype": "xfs", "label": "", "name": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "size": "4G", "type": "crypt", "uuid": "88582122-6df2-47e0-8975-fa2a6f19c3a0" }, "/dev/sda": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.218) 0:06:59.061 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003306", "end": "2023-11-23 00:29:26.486604", "rc": 0, "start": "2023-11-23 00:29:26.483298" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 /dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.217) 0:06:59.278 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003531", "end": "2023-11-23 00:29:26.699089", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:29:26.695558" } STDOUT: luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164 /dev/mapper/foo-test1 - TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.211) 0:06:59.489 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml for sut TASK [Set _storage_pool_tests] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:5 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.029) 0:06:59.519 ***** ok: [sut] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [Verify pool subset] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool.yml:18 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.013) 0:06:59.532 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:2 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.031) 0:06:59.564 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_count": "1", "_storage_test_pool_pvs_lvm": [ "/dev/sda" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:13 Thursday 23 November 2023 00:29:26 +0000 (0:00:00.020) 0:06:59.585 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda", "pv": "/dev/sda" } TASK [Set pvs lvm length] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:22 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.213) 0:06:59.798 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": "1" }, "changed": false } TASK [Set pool pvs] ************************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:27 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.019) 0:06:59.817 ***** ok: [sut] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:33 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.019) 0:06:59.836 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:42 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.019) 0:06:59.856 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:48 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.017) 0:06:59.873 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [Set expected pv type] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:54 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.017) 0:06:59.891 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:59 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:06:59.905 ***** ok: [sut] => (item=/dev/sda) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:73 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.058) 0:06:59.964 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml for sut TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:8 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.031) 0:06:59.995 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:14 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.010 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:21 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.024 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:28 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.015) 0:07:00.039 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md chunk size regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:35 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.053 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:45 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.068 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:54 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.082 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:64 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.096 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:74 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.013) 0:07:00.110 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:85 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.015) 0:07:00.125 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variables used by tests] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-md.yml:95 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.013) 0:07:00.139 ***** ok: [sut] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_chunk_size_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:76 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.013) 0:07:00.153 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml for sut TASK [Validate pool member LVM RAID settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-lvmraid.yml:2 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.029) 0:07:00.182 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml for sut TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:8 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.033) 0:07:00.216 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:16 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.015) 0:07:00.232 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:21 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.247 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV stripe size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:29 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.261 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested stripe size] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:34 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.015) 0:07:00.277 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected stripe size] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:40 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.292 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check stripe size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-lvmraid.yml:46 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.306 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:79 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.320 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml for sut TASK [Validate pool member thinpool settings] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-thin.yml:2 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.031) 0:07:00.351 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml for sut TASK [Get information about thinpool] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:8 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.028) 0:07:00.380 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:16 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.395 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:23 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.409 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Reset variable used by test] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-thin.yml:27 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.015) 0:07:00.424 ***** ok: [sut] => { "ansible_facts": { "storage_test_thin_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:82 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.439 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml for sut TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:5 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.034) 0:07:00.473 ***** ok: [sut] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:13 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.017) 0:07:00.491 ***** skipping: [sut] => (item=/dev/sda) => { "_storage_test_pool_member_path": "/dev/sda", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:20 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.018) 0:07:00.510 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml for sut TASK [Set variables used by tests] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:2 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.028) 0:07:00.538 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:9 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.018) 0:07:00.556 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:18 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.018) 0:07:00.574 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:27 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.589 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:37 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.015) 0:07:00.604 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-crypttab.yml:47 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.013) 0:07:00.618 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-encryption.yml:27 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.632 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:85 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.646 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml for sut TASK [Validate pool member VDO settings] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-members-vdo.yml:2 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.034) 0:07:00.680 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml for sut TASK [Get information about VDO deduplication] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:9 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.030) 0:07:00.711 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:16 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.014) 0:07:00.725 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:22 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.013) 0:07:00.739 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about VDO compression] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:28 Thursday 23 November 2023 00:29:27 +0000 (0:00:00.016) 0:07:00.756 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is off] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:35 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.015) 0:07:00.771 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check if VDO deduplication is on] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:41 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.016) 0:07:00.787 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-pool-member-vdo.yml:47 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.016) 0:07:00.804 ***** ok: [sut] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [Clean up test variables] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-members.yml:88 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.018) 0:07:00.823 ***** ok: [sut] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Verify the volumes] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-pool-volumes.yml:3 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.019) 0:07:00.843 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.080) 0:07:00.923 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.021) 0:07:00.944 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.076) 0:07:01.020 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.018) 0:07:01.039 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1012662, "block_size": 4096, "block_total": 1028096, "block_used": 15434, "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fstype": "xfs", "inode_available": 2088957, "inode_total": 2088960, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4147863552, "size_total": 4211081216, "uuid": "88582122-6df2-47e0-8975-fa2a6f19c3a0" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1012662, "block_size": 4096, "block_total": 1028096, "block_used": 15434, "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fstype": "xfs", "inode_available": 2088957, "inode_total": 2088960, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=32k,noquota", "size_available": 4147863552, "size_total": 4211081216, "uuid": "88582122-6df2-47e0-8975-fa2a6f19c3a0" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.024) 0:07:01.063 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.017) 0:07:01.080 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.019) 0:07:01.100 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.018) 0:07:01.118 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.016) 0:07:01.135 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.149 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.164 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.018) 0:07:01.182 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.197 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.211 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.016) 0:07:01.228 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.242 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.031) 0:07:01.273 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.017) 0:07:01.291 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.018) 0:07:01.310 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.017) 0:07:01.327 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.342 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.020) 0:07:01.362 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.020) 0:07:01.383 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699365.1721072, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699361.7472425, "dev": 5, "device_type": 64768, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1899, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699361.7472425, "nlink": 1, "path": "/dev/mapper/foo-test1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.217) 0:07:01.601 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.019) 0:07:01.620 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.014) 0:07:01.634 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.017) 0:07:01.652 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.016) 0:07:01.669 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.016) 0:07:01.685 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:29:28 +0000 (0:00:00.018) 0:07:01.704 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699365.1541078, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699361.9672337, "dev": 5, "device_type": 64769, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 2007, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1700699361.9672337, "nlink": 1, "path": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:29:29 +0000 (0:00:00.219) 0:07:01.923 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:29:31 +0000 (0:00:02.563) 0:07:04.487 ***** ok: [sut] => { "changed": false, "cmd": [ "cryptsetup", "luksDump", "/dev/mapper/foo-test1" ], "delta": "0:00:00.008181", "end": "2023-11-23 00:29:31.916423", "rc": 0, "start": "2023-11-23 00:29:31.908242" } STDOUT: LUKS header information Version: 2 Epoch: 3 Metadata area: 16384 [bytes] Keyslots area: 16744448 [bytes] UUID: 87ddbe6c-ff42-47dc-9b03-3e5b4e379164 Label: (no label) Subsystem: (no subsystem) Flags: (no flags) Data segments: 0: crypt offset: 16777216 [bytes] length: (whole device) cipher: aes-xts-plain64 sector: 512 [bytes] Keyslots: 0: luks2 Key: 512 bits Priority: normal Cipher: aes-xts-plain64 Cipher key: 512 bits PBKDF: argon2id Time cost: 4 Memory: 717298 Threads: 2 Salt: 28 b8 4d b9 8b ad fd 6d da 20 91 09 57 d0 98 b2 93 a3 2b 77 98 e2 8f 57 1b 83 c6 07 2b d3 4d b6 AF stripes: 4000 AF hash: sha256 Area offset:32768 [bytes] Area length:258048 [bytes] Digest ID: 0 Tokens: Digests: 0: pbkdf2 Hash: sha256 Iterations: 93756 Salt: f1 fd 64 15 0d a9 32 6e 78 de c3 28 3a bb ef 9c cc 4d 21 2a d5 57 31 9c d5 b8 e7 01 5c 7b 14 40 Digest: 45 a5 51 33 19 6c eb 10 1b f2 1b 58 35 bd 09 e4 3d 82 34 ed fb bd ed dc ab 30 00 a9 05 b1 c6 87 TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:29:31 +0000 (0:00:00.219) 0:07:04.706 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:29:31 +0000 (0:00:00.019) 0:07:04.726 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:29:31 +0000 (0:00:00.021) 0:07:04.747 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.019) 0:07:04.767 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.018) 0:07:04.786 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.014) 0:07:04.801 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.015) 0:07:04.816 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.015) 0:07:04.831 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [ "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164 /dev/mapper/foo-test1 -" ], "_storage_test_expected_crypttab_entries": "1", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.020) 0:07:04.852 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.017) 0:07:04.870 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.018) 0:07:04.888 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.018) 0:07:04.907 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.018) 0:07:04.925 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:04.939 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.015) 0:07:04.954 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:04.968 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.014) 0:07:04.982 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:04.996 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:05.010 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:05.023 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.043) 0:07:05.066 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.015) 0:07:05.082 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:05.096 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.013) 0:07:05.110 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.211) 0:07:05.322 ***** ok: [sut] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.214) 0:07:05.537 ***** ok: [sut] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.022) 0:07:05.559 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:29:32 +0000 (0:00:00.016) 0:07:05.575 ***** ok: [sut] => { "bytes": 10726680821, "changed": false, "lvm": "9g", "parted": "9GiB", "size": "9 GiB" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.211) 0:07:05.787 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.018) 0:07:05.805 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.018) 0:07:05.824 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.017) 0:07:05.841 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.017) 0:07:05.859 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:05.873 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.015) 0:07:05.888 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:05.903 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.013) 0:07:05.917 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:05.931 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:05.945 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:05.959 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.016) 0:07:05.976 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:05.990 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:06.005 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:06.020 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.014) 0:07:06.034 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.015) 0:07:06.049 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.020) 0:07:06.069 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.015) 0:07:06.084 ***** ok: [sut] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.021) 0:07:06.106 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.020) 0:07:06.127 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.021) 0:07:06.149 ***** ok: [sut] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "foo/test1" ], "delta": "0:00:00.032598", "end": "2023-11-23 00:29:33.601921", "rc": 0, "start": "2023-11-23 00:29:33.569323" } STDOUT: LVM2_LV_NAME=test1 LVM2_LV_ATTR=-wi-ao---- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=linear TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.243) 0:07:06.392 ***** ok: [sut] => { "ansible_facts": { "storage_test_lv_segtype": [ "linear" ] }, "changed": false } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.020) 0:07:06.413 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.021) 0:07:06.434 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.018) 0:07:06.452 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.016) 0:07:06.468 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.016) 0:07:06.484 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.016) 0:07:06.501 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.012) 0:07:06.514 ***** TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.012) 0:07:06.526 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Clean up] **************************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:493 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.015) 0:07:06.542 ***** TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.050) 0:07:06.592 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for sut TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.023) 0:07:06.615 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.018) 0:07:06.634 ***** skipping: [sut] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [sut] => (item=Fedora.yml) => { "ansible_facts": { "_storage_copr_packages": [ { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" } ], "_storage_copr_support_packages": [ "dnf-plugins-core" ], "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap" ] }, "ansible_included_var_files": [ "/WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/vars/Fedora.yml" ], "ansible_loop_var": "item", "changed": false, "item": "Fedora.yml" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } skipping: [sut] => (item=Fedora_38.yml) => { "ansible_loop_var": "item", "changed": false, "item": "Fedora_38.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Check if system is ostree] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:26 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.038) 0:07:06.672 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Set flag to indicate system is ostree] ****** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:31 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.015) 0:07:06.688 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Define an empty list of pools to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.049) 0:07:06.738 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Define an empty list of volumes to be used in testing] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 23 November 2023 00:29:33 +0000 (0:00:00.015) 0:07:06.753 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : Include the appropriate provider tasks] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 23 November 2023 00:29:34 +0000 (0:00:00.014) 0:07:06.768 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for sut TASK [linux-system-roles.storage : Make sure blivet is available] ************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 23 November 2023 00:29:34 +0000 (0:00:00.033) 0:07:06.802 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Show storage_pools] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:11 Thursday 23 November 2023 00:29:36 +0000 (0:00:02.500) 0:07:09.302 ***** ok: [sut] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : Show storage_volumes] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:16 Thursday 23 November 2023 00:29:36 +0000 (0:00:00.017) 0:07:09.320 ***** ok: [sut] => { "storage_volumes": [ { "disks": [ "sda" ], "name": "foo", "state": "absent", "type": "disk" } ] } TASK [linux-system-roles.storage : Get required packages] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:21 Thursday 23 November 2023 00:29:36 +0000 (0:00:00.018) 0:07:09.338 ***** ok: [sut] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Enable copr repositories if needed] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:34 Thursday 23 November 2023 00:29:38 +0000 (0:00:02.184) 0:07:11.522 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for sut TASK [linux-system-roles.storage : Check if the COPR support packages should be installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 23 November 2023 00:29:38 +0000 (0:00:00.027) 0:07:11.550 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure COPR support packages are present] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 23 November 2023 00:29:38 +0000 (0:00:00.022) 0:07:11.572 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Enable COPRs] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:20 Thursday 23 November 2023 00:29:38 +0000 (0:00:00.014) 0:07:11.586 ***** skipping: [sut] => (item={'repository': 'rhawalsh/dm-vdo', 'packages': ['vdo', 'kmod-vdo']}) => { "ansible_loop_var": "repo", "changed": false, "repo": { "packages": [ "vdo", "kmod-vdo" ], "repository": "rhawalsh/dm-vdo" }, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Make sure required packages are installed] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 Thursday 23 November 2023 00:29:38 +0000 (0:00:00.024) 0:07:11.611 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : Get service facts] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:57 Thursday 23 November 2023 00:29:41 +0000 (0:00:02.464) 0:07:14.076 ***** ok: [sut] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "arp-ethers.service": { "name": "arp-ethers.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "bluetooth.service": { "name": "bluetooth.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd-restricted.service": { "name": "chronyd-restricted.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "cloud-init-hotplugd.service": { "name": "cloud-init-hotplugd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.bluez.service": { "name": "dbus-org.bluez.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.oom1.service": { "name": "dbus-org.freedesktop.oom1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.portable1.service": { "name": "dbus-org.freedesktop.portable1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.resolve1.service": { "name": "dbus-org.freedesktop.resolve1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-system-upgrade-cleanup.service": { "name": "dnf-system-upgrade-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "dnf-system-upgrade.service": { "name": "dnf-system-upgrade.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "firewalld.service": { "name": "firewalld.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fsidd.service": { "name": "fsidd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "stopped", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "unknown" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "import-state.service": { "name": "import-state.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "loadmodules.service": { "name": "loadmodules.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@dm_mod.service": { "name": "modprobe@dm_mod.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "modprobe@loop.service": { "name": "modprobe@loop.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nftables.service": { "name": "nftables.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "pcscd.service": { "name": "pcscd.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "plymouth-halt.service": { "name": "plymouth-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-kexec.service": { "name": "plymouth-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-poweroff.service": { "name": "plymouth-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-quit.service": { "name": "plymouth-quit.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-read-write.service": { "name": "plymouth-read-write.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-reboot.service": { "name": "plymouth-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "static" }, "plymouth-switch-root-initramfs.service": { "name": "plymouth-switch-root-initramfs.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-switch-root.service": { "name": "plymouth-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "stopped", "status": "static" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "restraintd.service": { "name": "restraintd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rngd.service": { "name": "rngd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "rpmdb-migrate.service": { "name": "rpmdb-migrate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "selinux-check-proper-disable.service": { "name": "selinux-check-proper-disable.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "ssh-host-keys-migration.service": { "name": "ssh-host-keys-migration.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-plymouth.service": { "name": "systemd-ask-password-plymouth.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-bless-boot.service": { "name": "systemd-bless-boot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-boot-random-seed.service": { "name": "systemd-boot-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-update.service": { "name": "systemd-boot-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-growfs-root.service": { "name": "systemd-growfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-growfs@.service": { "name": "systemd-growfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-homed-activate.service": { "name": "systemd-homed-activate.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-homed.service": { "name": "systemd-homed.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-oomd.service": { "name": "systemd-oomd.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-pcrfs-root.service": { "name": "systemd-pcrfs-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pcrfs@.service": { "name": "systemd-pcrfs@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-pcrmachine.service": { "name": "systemd-pcrmachine.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-initrd.service": { "name": "systemd-pcrphase-initrd.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase-sysinit.service": { "name": "systemd-pcrphase-sysinit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-pcrphase.service": { "name": "systemd-pcrphase.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-portabled.service": { "name": "systemd-portabled.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-resolved.service": { "name": "systemd-resolved.service", "source": "systemd", "state": "running", "status": "enabled" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-sysupdate-reboot.service": { "name": "systemd-sysupdate-reboot.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysupdate.service": { "name": "systemd-sysupdate.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-time-wait-sync.service": { "name": "systemd-time-wait-sync.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-userdbd.service": { "name": "systemd-userdbd.service", "source": "systemd", "state": "running", "status": "indirect" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-zram-setup@.service": { "name": "systemd-zram-setup@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-zram-setup@zram0.service": { "name": "systemd-zram-setup@zram0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "target.service": { "name": "target.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "targetclid.service": { "name": "targetclid.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "udisks2.service": { "name": "udisks2.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "unknown" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "unknown" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:64 Thursday 23 November 2023 00:29:43 +0000 (0:00:02.368) 0:07:16.445 ***** ok: [sut] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:78 Thursday 23 November 2023 00:29:43 +0000 (0:00:00.028) 0:07:16.473 ***** TASK [linux-system-roles.storage : Manage the pools and volumes to match the specified state] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Thursday 23 November 2023 00:29:43 +0000 (0:00:00.014) 0:07:16.487 ***** changed: [sut] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [ { "backing_device": "/dev/mapper/foo-test1", "name": "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "password": "-", "state": "absent" } ], "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/sda", "_mount_id": "UUID=FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7", "_raw_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:98 Thursday 23 November 2023 00:29:46 +0000 (0:00:03.136) 0:07:19.624 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:110 Thursday 23 November 2023 00:29:46 +0000 (0:00:00.015) 0:07:19.639 ***** TASK [linux-system-roles.storage : Show blivet_output] ************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:116 Thursday 23 November 2023 00:29:46 +0000 (0:00:00.013) 0:07:19.652 ***** ok: [sut] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "fs_type": null }, { "action": "destroy format", "device": "/dev/mapper/foo-test1", "fs_type": "luks" }, { "action": "destroy device", "device": "/dev/mapper/foo-test1", "fs_type": null }, { "action": "destroy device", "device": "/dev/foo", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "lvmpv" } ], "changed": true, "crypts": [ { "backing_device": "/dev/mapper/foo-test1", "name": "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "password": "-", "state": "absent" } ], "failed": false, "leaves": [ "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/sdd", "/dev/sde", "/dev/sdf", "/dev/sdg", "/dev/sdh", "/dev/sdi", "/dev/xvda1", "/dev/xvda2", "/dev/zram0" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "absent" } ], "packages": [ "e2fsprogs" ], "pools": [], "volumes": [ { "_device": "/dev/sda", "_mount_id": "UUID=FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7", "_raw_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } } TASK [linux-system-roles.storage : Set the list of pools for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:121 Thursday 23 November 2023 00:29:46 +0000 (0:00:00.018) 0:07:19.670 ***** ok: [sut] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : Set the list of volumes for test verification] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:125 Thursday 23 November 2023 00:29:46 +0000 (0:00:00.016) 0:07:19.687 ***** ok: [sut] => { "ansible_facts": { "_storage_volumes_list": [ { "_device": "/dev/sda", "_mount_id": "UUID=FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7", "_raw_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] }, "changed": false } TASK [linux-system-roles.storage : Remove obsolete mounts] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:141 Thursday 23 November 2023 00:29:46 +0000 (0:00:00.017) 0:07:19.704 ***** changed: [sut] => (item={'src': '/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164', 'path': '/opt/test1', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test1", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164" } TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:153 Thursday 23 November 2023 00:29:47 +0000 (0:00:00.229) 0:07:19.934 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Set up new/current mounts] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:158 Thursday 23 November 2023 00:29:48 +0000 (0:00:00.996) 0:07:20.931 ***** TASK [linux-system-roles.storage : Manage mount ownership/permissions] ********* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:170 Thursday 23 November 2023 00:29:48 +0000 (0:00:00.016) 0:07:20.947 ***** TASK [linux-system-roles.storage : Tell systemd to refresh its view of /etc/fstab] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:185 Thursday 23 November 2023 00:29:48 +0000 (0:00:00.014) 0:07:20.961 ***** ok: [sut] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : Retrieve facts for the /etc/crypttab file] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:193 Thursday 23 November 2023 00:29:49 +0000 (0:00:00.998) 0:07:21.960 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699365.1701071, "attr_flags": "e", "attributes": [ "extents" ], "block_size": 4096, "blocks": 8, "charset": "us-ascii", "checksum": "27365370d567fd555eb6009b10f9aefa950259ee", "ctime": 1700699365.150108, "dev": 51714, "device_type": 0, "executable": false, "exists": true, "gid": 0, "gr_name": "root", "inode": 263319, "isblk": false, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": true, "issock": false, "isuid": false, "mimetype": "text/plain", "mode": "0600", "mtime": 1700699365.149108, "nlink": 1, "path": "/etc/crypttab", "pw_name": "root", "readable": true, "rgrp": false, "roth": false, "rusr": true, "size": 66, "uid": 0, "version": "767875063", "wgrp": false, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [linux-system-roles.storage : Manage /etc/crypttab to account for changes we just made] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 23 November 2023 00:29:49 +0000 (0:00:00.227) 0:07:22.187 ***** changed: [sut] => (item={'backing_device': '/dev/mapper/foo-test1', 'name': 'luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164', 'password': '-', 'state': 'absent'}) => { "ansible_loop_var": "entry", "backup": "", "changed": true, "entry": { "backing_device": "/dev/mapper/foo-test1", "name": "luks-87ddbe6c-ff42-47dc-9b03-3e5b4e379164", "password": "-", "state": "absent" }, "found": 1 } MSG: 1 line(s) removed TASK [linux-system-roles.storage : Update facts] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:220 Thursday 23 November 2023 00:29:49 +0000 (0:00:00.233) 0:07:22.421 ***** ok: [sut] TASK [Verify role results] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:503 Thursday 23 November 2023 00:29:50 +0000 (0:00:00.805) 0:07:23.226 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml for sut TASK [Print out pool information] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:2 Thursday 23 November 2023 00:29:50 +0000 (0:00:00.034) 0:07:23.260 ***** skipping: [sut] => {} TASK [Print out volume information] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:7 Thursday 23 November 2023 00:29:50 +0000 (0:00:00.014) 0:07:23.274 ***** ok: [sut] => { "_storage_volumes_list": [ { "_device": "/dev/sda", "_mount_id": "UUID=FtoydN-sBSX-hgPs-69MT-FBM0-dI92-gs16W7", "_raw_device": "/dev/sda", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "lvmpv", "mount_check": 0, "mount_device_identifier": "uuid", "mount_group": null, "mount_mode": null, "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "mount_user": null, "name": "foo", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "raid_stripe_size": null, "size": 10737418240, "state": "absent", "thin": null, "thin_pool_name": null, "thin_pool_size": null, "type": "disk", "vdo_pool_size": null } ] } TASK [Collect info about the volumes.] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:15 Thursday 23 November 2023 00:29:50 +0000 (0:00:00.017) 0:07:23.292 ***** ok: [sut] => { "changed": false, "info": { "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdd": { "fstype": "", "label": "", "name": "/dev/sdd", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sde": { "fstype": "", "label": "", "name": "/dev/sde", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdf": { "fstype": "", "label": "", "name": "/dev/sdf", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdg": { "fstype": "", "label": "", "name": "/dev/sdg", "size": "1T", "type": "disk", "uuid": "" }, "/dev/sdh": { "fstype": "", "label": "", "name": "/dev/sdh", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdi": { "fstype": "", "label": "", "name": "/dev/sdi", "size": "10G", "type": "disk", "uuid": "" }, "/dev/xvda": { "fstype": "", "label": "", "name": "/dev/xvda", "size": "250G", "type": "disk", "uuid": "" }, "/dev/xvda1": { "fstype": "", "label": "", "name": "/dev/xvda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/xvda2": { "fstype": "ext4", "label": "", "name": "/dev/xvda2", "size": "250G", "type": "partition", "uuid": "702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c" }, "/dev/zram0": { "fstype": "", "label": "", "name": "/dev/zram0", "size": "3.6G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:20 Thursday 23 November 2023 00:29:50 +0000 (0:00:00.218) 0:07:23.510 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003358", "end": "2023-11-23 00:29:50.932615", "rc": 0, "start": "2023-11-23 00:29:50.929257" } STDOUT: # # /etc/fstab # Created by anaconda on Wed Nov 15 19:19:52 2023 # # Accessible filesystems, by reference, are maintained under '/dev/disk/'. # See man pages fstab(5), findfs(8), mount(8) and/or blkid(8) for more info. # # After editing this file, run 'systemctl daemon-reload' to update systemd # units generated from this file. # UUID=702a114d-b3c0-4c4e-9150-1bd3c1b2ac8c / ext4 defaults 1 1 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat /mnt/redhat nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/qa /mnt/qa nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 vtap-eng01.storage.rdu2.redhat.com:/vol/engarchive /mnt/engarchive nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 nest.test.redhat.com:/mnt/tpsdist /mnt/tpsdist nfs defaults,rsize=8192,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_engineering_nfs/devarchive/redhat/brewroot /mnt/brew nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 ntap-bos-c01-eng01-nfs01b.storage.bos.redhat.com:/devops_brew_scratch_nfs/scratch /mnt/brew_scratch nfs ro,rsize=32768,wsize=8192,bg,noauto,noatime,nosuid,nodev,intr 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:25 Thursday 23 November 2023 00:29:50 +0000 (0:00:00.212) 0:07:23.723 ***** ok: [sut] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.003344", "end": "2023-11-23 00:29:51.147184", "failed_when_result": false, "rc": 0, "start": "2023-11-23 00:29:51.143840" } TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:34 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.213) 0:07:23.937 ***** TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:44 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:23.950 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml for sut TASK [Set storage volume test variables] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:2 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.027) 0:07:23.978 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [Run test verify for {{ storage_test_volume_subset }}] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:21 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.017) 0:07:23.996 ***** included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml for sut included: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml for sut TASK [Get expected mount device based on device type] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:7 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.068) 0:07:24.064 ***** ok: [sut] => { "ansible_facts": { "storage_test_device_path": "/dev/sda" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:16 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.017) 0:07:24.082 ***** ok: [sut] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Get information about the mountpoint directory] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:38 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.021) 0:07:24.104 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by device] ******************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:51 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.119 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:63 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.133 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount directory user] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:71 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.017) 0:07:24.150 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory group] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:83 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.165 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify mount directory permissions] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:95 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.015) 0:07:24.181 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the mount fs type] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:110 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.195 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get path of test volume device] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:122 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.015) 0:07:24.210 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:128 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.224 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:134 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.238 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-mount.yml:146 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.252 ***** ok: [sut] => { "ansible_facts": { "storage_test_found_mount_stat": null, "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:2 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.266 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:40 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.030) 0:07:24.297 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:48 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.045) 0:07:24.342 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:58 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.019) 0:07:24.362 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fstab.yml:71 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.376 ***** ok: [sut] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:3 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.014) 0:07:24.390 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-fs.yml:12 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.404 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:3 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.417 ***** ok: [sut] => { "changed": false, "stat": { "atime": 1700699386.7372553, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1700699386.7372553, "dev": 5, "device_type": 2048, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 574, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/blockdevice", "mode": "0660", "mtime": 1700699386.7372553, "nlink": 1, "path": "/dev/sda", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:9 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.219) 0:07:24.637 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Verify the presence/absence of the device node] ************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:16 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.019) 0:07:24.656 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about this volume] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:24 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.016) 0:07:24.673 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Process volume type (set initial value) (1/2)] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:30 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.687 ***** ok: [sut] => { "ansible_facts": { "st_volume_type": "disk" }, "changed": false } TASK [Process volume type (get RAID value) (2/2)] ****************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:34 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.016) 0:07:24.703 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-device.yml:39 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.015) 0:07:24.718 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:3 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.011) 0:07:24.730 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Thursday 23 November 2023 00:29:51 +0000 (0:00:00.013) 0:07:24.744 ***** ok: [sut] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:17 Thursday 23 November 2023 00:29:54 +0000 (0:00:02.459) 0:07:27.204 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:23 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.218 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:32 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.233 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:45 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.012) 0:07:27.245 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:51 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.015) 0:07:27.260 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:56 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.275 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:69 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.011) 0:07:27.287 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:81 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.011) 0:07:27.298 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set test variables] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:94 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.011) 0:07:27.310 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:106 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.019) 0:07:27.330 ***** ok: [sut] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:114 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.018) 0:07:27.349 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:122 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.363 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:131 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.376 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clear test variables] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:140 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.390 ***** ok: [sut] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [Get information about RAID] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:8 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.403 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set active devices regex] ************************************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:14 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.417 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set spare devices regex] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:21 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.015) 0:07:27.433 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set md version regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:28 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.447 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set chunk size regex] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:35 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.461 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the chunk size] **************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:45 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.475 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID active devices count] ***************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:54 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.489 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID spare devices count] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:63 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.503 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID metadata version] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:72 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.017) 0:07:27.520 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check RAID chunk size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-md.yml:81 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.015) 0:07:27.536 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the actual size of the volume] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:3 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.550 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested size of the volume] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:11 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.565 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:20 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.579 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:28 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.593 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Get the size of parent/pool device] ************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:32 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.016) 0:07:27.610 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show test pool] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:46 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.624 ***** skipping: [sut] => {} TASK [Show test blockinfo] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:50 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.639 ***** skipping: [sut] => {} TASK [Show test pool size] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:54 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.653 ***** skipping: [sut] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:58 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.667 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default thin pool reserved space values] ********************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:68 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.680 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default minimal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:72 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.015) 0:07:27.696 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Default maximal thin pool reserved space size] *************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:77 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.710 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate maximum usable space in thin pool] ***************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:83 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.724 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply upper size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:88 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.014) 0:07:27.739 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Apply lower size limit to max usable thin pool space] ******************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:96 Thursday 23 November 2023 00:29:54 +0000 (0:00:00.013) 0:07:27.753 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Convert maximum usable thin pool space from int to Size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:104 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:27.767 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show max thin pool size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:109 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.015) 0:07:27.783 ***** skipping: [sut] => {} TASK [Show volume thin pool size] ********************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:113 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:27.797 ***** skipping: [sut] => {} TASK [Show test volume size] *************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:117 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:27.812 ***** skipping: [sut] => {} TASK [Establish base value for expected thin pool size] ************************ task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:121 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:27.826 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:129 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.013) 0:07:27.840 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected thin pool volume size] ***************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:138 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.013) 0:07:27.854 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Calculate the expected thin pool volume size based on percentage value] *** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:142 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.015) 0:07:27.869 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Replace expected volume size with calculated value] ********************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:150 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:27.883 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Show actual size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:156 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.013) 0:07:27.897 ***** ok: [sut] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [Show expected size] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:160 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.016) 0:07:27.914 ***** ok: [sut] => { "storage_test_expected_size": "4294967296" } TASK [Assert expected size is actual size] ************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-size.yml:164 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.015) 0:07:27.929 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:5 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.013) 0:07:27.943 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV segment type] ***************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:13 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.015) 0:07:27.958 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check segment type] ****************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:18 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.013) 0:07:27.972 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set LV cache size] ******************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:27 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:27.986 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Parse the requested cache size] ****************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:35 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:28.000 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Set expected cache size] ************************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:41 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.015) 0:07:28.015 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-cache.yml:47 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:28.030 ***** skipping: [sut] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume.yml:27 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.015) 0:07:28.045 ***** ok: [sut] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Clean up variable namespace] ********************************************* task path: /WORKDIR/git-refactor-ostreegqzakvvd/tests/verify-role-results.yml:54 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.014) 0:07:28.059 ***** ok: [sut] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* sut : ok=1183 changed=62 unreachable=0 failed=9 skipped=1014 rescued=9 ignored=0 Thursday 23 November 2023 00:29:55 +0000 (0:00:00.008) 0:07:28.068 ***** =============================================================================== linux-system-roles.storage : Manage the pools and volumes to match the specified state -- 11.62s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Manage the pools and volumes to match the specified state -- 11.22s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Manage the pools and volumes to match the specified state -- 11.16s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Manage the pools and volumes to match the specified state -- 11.05s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Manage the pools and volumes to match the specified state -- 10.86s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 9.42s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Make sure blivet is available -------------- 6.20s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 linux-system-roles.storage : Make sure required packages are installed --- 3.51s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 3.14s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 3.07s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Gathering Facts --------------------------------------------------------- 2.78s /WORKDIR/git-refactor-ostreegqzakvvd/tests/tests_luks.yml:2 ------------------- Ensure cryptsetup is present -------------------------------------------- 2.75s /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 linux-system-roles.storage : Manage the pools and volumes to match the specified state --- 2.69s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:84 Ensure cryptsetup is present -------------------------------------------- 2.62s /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 linux-system-roles.storage : Make sure required packages are installed --- 2.59s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:41 linux-system-roles.storage : Make sure blivet is available -------------- 2.58s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 linux-system-roles.storage : Make sure blivet is available -------------- 2.56s /WORKDIR/git-refactor-ostreegqzakvvd/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Ensure cryptsetup is present -------------------------------------------- 2.56s /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.56s /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 Ensure cryptsetup is present -------------------------------------------- 2.56s /WORKDIR/git-refactor-ostreegqzakvvd/tests/test-verify-volume-encryption.yml:10 ---^---^---^---^---^--- # STDERR: ---v---v---v---v---v--- ---^---^---^---^---^---