ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmpi19f9hzy executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_create_thinp_then_remove.yml *********************************** 1 plays in /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:2 Thursday 21 July 2022 08:11:56 +0000 (0:00:00.034) 0:00:00.034 ********* ok: [/cache/rhel-9.qcow2.snap] META: ran handlers TASK [include_role : linux-system-roles.storage] ******************************* task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:14 Thursday 21 July 2022 08:11:57 +0000 (0:00:01.373) 0:00:01.407 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:11:57 +0000 (0:00:00.035) 0:00:01.443 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:11:57 +0000 (0:00:00.031) 0:00:01.474 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:11:58 +0000 (0:00:00.530) 0:00:02.005 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:11:58 +0000 (0:00:00.055) 0:00:02.060 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:11:58 +0000 (0:00:00.030) 0:00:02.090 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:11:58 +0000 (0:00:00.030) 0:00:02.120 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:11:58 +0000 (0:00:00.045) 0:00:02.166 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:11:58 +0000 (0:00:00.018) 0:00:02.184 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:11:59 +0000 (0:00:01.258) 0:00:03.443 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:11:59 +0000 (0:00:00.034) 0:00:03.477 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:12:00 +0000 (0:00:00.033) 0:00:03.511 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:12:00 +0000 (0:00:00.701) 0:00:04.213 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:12:00 +0000 (0:00:00.041) 0:00:04.254 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:12:00 +0000 (0:00:00.030) 0:00:04.285 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:12:00 +0000 (0:00:00.034) 0:00:04.319 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:12:00 +0000 (0:00:00.030) 0:00:04.349 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:12:01 +0000 (0:00:00.832) 0:00:05.182 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:12:03 +0000 (0:00:01.802) 0:00:06.984 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:12:03 +0000 (0:00:00.056) 0:00:07.041 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:12:03 +0000 (0:00:00.021) 0:00:07.062 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.522) 0:00:07.585 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.034) 0:00:07.619 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.019) 0:00:07.639 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.033) 0:00:07.672 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.033) 0:00:07.706 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.034) 0:00:07.741 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.032) 0:00:07.773 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.021) 0:00:07.795 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.033) 0:00:07.829 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.024) 0:00:07.853 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.481) 0:00:08.335 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:12:04 +0000 (0:00:00.021) 0:00:08.356 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:17 Thursday 21 July 2022 08:12:05 +0000 (0:00:01.021) 0:00:09.377 ********* included: /tmp/tmpsvhdx31t/tests/get_unused_disk.yml for /cache/rhel-9.qcow2.snap TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmpsvhdx31t/tests/get_unused_disk.yml:2 Thursday 21 July 2022 08:12:05 +0000 (0:00:00.055) 0:00:09.433 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "disks": [ "sda", "sdb", "sdc" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmpsvhdx31t/tests/get_unused_disk.yml:9 Thursday 21 July 2022 08:12:06 +0000 (0:00:00.483) 0:00:09.917 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "unused_disks": [ "sda", "sdb", "sdc" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmpsvhdx31t/tests/get_unused_disk.yml:14 Thursday 21 July 2022 08:12:06 +0000 (0:00:00.037) 0:00:09.954 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/get_unused_disk.yml:19 Thursday 21 July 2022 08:12:06 +0000 (0:00:00.036) 0:00:09.991 ********* ok: [/cache/rhel-9.qcow2.snap] => { "unused_disks": [ "sda", "sdb", "sdc" ] } TASK [Create a thinpool device] ************************************************ task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:21 Thursday 21 July 2022 08:12:06 +0000 (0:00:00.032) 0:00:10.023 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:12:06 +0000 (0:00:00.041) 0:00:10.065 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:12:06 +0000 (0:00:00.031) 0:00:10.097 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:12:07 +0000 (0:00:00.515) 0:00:10.612 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:12:07 +0000 (0:00:00.058) 0:00:10.671 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:12:07 +0000 (0:00:00.030) 0:00:10.701 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:12:07 +0000 (0:00:00.031) 0:00:10.733 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:12:07 +0000 (0:00:00.043) 0:00:10.777 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:12:07 +0000 (0:00:00.019) 0:00:10.796 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:12:08 +0000 (0:00:00.903) 0:00:11.700 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "state": "present", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "3g", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:12:08 +0000 (0:00:00.065) 0:00:11.766 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:12:08 +0000 (0:00:00.063) 0:00:11.829 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:12:09 +0000 (0:00:01.374) 0:00:13.204 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:12:09 +0000 (0:00:00.066) 0:00:13.270 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:12:09 +0000 (0:00:00.032) 0:00:13.303 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:12:09 +0000 (0:00:00.035) 0:00:13.338 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:12:09 +0000 (0:00:00.030) 0:00:13.368 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:12:10 +0000 (0:00:00.858) 0:00:14.227 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:12:12 +0000 (0:00:01.751) 0:00:15.979 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:12:12 +0000 (0:00:00.059) 0:00:16.038 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:12:12 +0000 (0:00:00.023) 0:00:16.062 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "create format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdc1", "fs_type": null }, { "action": "create format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "lvm2", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:12:17 +0000 (0:00:04.538) 0:00:20.601 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:12:17 +0000 (0:00:00.037) 0:00:20.639 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:12:17 +0000 (0:00:00.022) 0:00:20.661 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdc1", "fs_type": null }, { "action": "create format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "lvm2", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:12:17 +0000 (0:00:00.039) 0:00:20.700 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:12:17 +0000 (0:00:00.037) 0:00:20.738 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:12:17 +0000 (0:00:00.035) 0:00:20.773 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:12:17 +0000 (0:00:00.046) 0:00:20.820 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:12:18 +0000 (0:00:00.952) 0:00:21.772 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:12:18 +0000 (0:00:00.552) 0:00:22.324 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:12:19 +0000 (0:00:00.700) 0:00:23.024 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:12:19 +0000 (0:00:00.415) 0:00:23.440 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:12:19 +0000 (0:00:00.022) 0:00:23.463 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:38 Thursday 21 July 2022 08:12:21 +0000 (0:00:01.091) 0:00:24.554 ********* included: /tmp/tmpsvhdx31t/tests/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:1 Thursday 21 July 2022 08:12:21 +0000 (0:00:00.042) 0:00:24.597 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:6 Thursday 21 July 2022 08:12:21 +0000 (0:00:00.048) 0:00:24.646 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:14 Thursday 21 July 2022 08:12:21 +0000 (0:00:00.035) 0:00:24.681 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "jHWuRa-IE9u-VNaw-TG3b-eUEu-7c97-hAZTqb" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "CvMQQK-EGRf-g62R-VzcS-QyRJ-JQlf-3f0oiq" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "39QUtr-6fIa-Np89-vDzq-yFkf-XlPB-dklI1i" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-08-11-44-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:19 Thursday 21 July 2022 08:12:21 +0000 (0:00:00.487) 0:00:25.168 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002816", "end": "2022-07-21 04:12:22.447010", "rc": 0, "start": "2022-07-21 04:12:22.444194" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/vg1-lv1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 Thursday 21 July 2022 08:12:22 +0000 (0:00:00.533) 0:00:25.702 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.003784", "end": "2022-07-21 04:12:23.822620", "failed_when_result": false, "rc": 1, "start": "2022-07-21 04:12:22.818836" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:33 Thursday 21 July 2022 08:12:23 +0000 (0:00:01.371) 0:00:27.073 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:5 Thursday 21 July 2022 08:12:23 +0000 (0:00:00.061) 0:00:27.134 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:18 Thursday 21 July 2022 08:12:23 +0000 (0:00:00.037) 0:00:27.172 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 08:12:23 +0000 (0:00:00.049) 0:00:27.221 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 08:12:23 +0000 (0:00:00.065) 0:00:27.287 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 08:12:25 +0000 (0:00:01.247) 0:00:28.534 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.051) 0:00:28.586 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.052) 0:00:28.638 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.054) 0:00:28.692 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.037) 0:00:28.730 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.051) 0:00:28.782 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.025) 0:00:28.807 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.069) 0:00:28.876 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:6 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.044) 0:00:28.920 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:12 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.026) 0:00:28.947 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:16 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.026) 0:00:28.974 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:20 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.024) 0:00:28.998 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:24 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.026) 0:00:29.024 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:30 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.026) 0:00:29.051 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:36 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.024) 0:00:29.076 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:44 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.034) 0:00:29.111 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.035) 0:00:29.146 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.046) 0:00:29.193 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.055) 0:00:29.248 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.028) 0:00:29.277 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.029) 0:00:29.306 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.029) 0:00:29.335 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.048) 0:00:29.384 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 08:12:25 +0000 (0:00:00.046) 0:00:29.430 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv1&&segtype=thin", "vg1" ], "delta": "0:00:00.034224", "end": "2022-07-21 04:12:26.582903", "rc": 0, "start": "2022-07-21 04:12:26.548679" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.403) 0:00:29.834 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.060) 0:00:29.895 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.064) 0:00:29.959 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.042) 0:00:30.001 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.050) 0:00:30.052 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.081) 0:00:30.133 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.032) 0:00:30.166 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.055) 0:00:30.221 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.050) 0:00:30.271 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.051) 0:00:30.322 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.039) 0:00:30.362 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.037) 0:00:30.400 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:12:26 +0000 (0:00:00.037) 0:00:30.437 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.041) 0:00:30.479 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.054) 0:00:30.534 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.050) 0:00:30.584 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.039) 0:00:30.623 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.036) 0:00:30.660 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.036) 0:00:30.696 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.035) 0:00:30.732 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.060) 0:00:30.792 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.055) 0:00:30.848 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.041) 0:00:30.890 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.038) 0:00:30.928 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.038) 0:00:30.966 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.038) 0:00:31.005 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.037) 0:00:31.042 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.053) 0:00:31.095 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.048) 0:00:31.143 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.025) 0:00:31.169 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.024) 0:00:31.193 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.023) 0:00:31.217 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.023) 0:00:31.241 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.025) 0:00:31.266 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.023) 0:00:31.290 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.023) 0:00:31.313 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.083) 0:00:31.397 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 08:12:27 +0000 (0:00:00.078) 0:00:31.476 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:2 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.045) 0:00:31.521 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:10 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.052) 0:00:31.574 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.080) 0:00:31.654 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.041) 0:00:31.695 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.058) 0:00:31.754 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.052) 0:00:31.807 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.055) 0:00:31.863 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.052) 0:00:31.915 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.027) 0:00:31.942 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.025) 0:00:31.968 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.025) 0:00:31.993 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.039) 0:00:32.033 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/vg1-lv1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.063) 0:00:32.096 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.057) 0:00:32.154 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.059) 0:00:32.213 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.039) 0:00:32.252 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.037) 0:00:32.289 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.042) 0:00:32.332 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 08:12:28 +0000 (0:00:00.040) 0:00:32.372 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658391137.2726676, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658391137.2726676, "dev": 5, "device_type": 64772, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1047, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658391137.2726676, "nlink": 1, "path": "/dev/mapper/vg1-lv1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.375) 0:00:32.747 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.037) 0:00:32.785 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.040) 0:00:32.826 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.037) 0:00:32.864 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.024) 0:00:32.888 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.039) 0:00:32.927 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 08:12:29 +0000 (0:00:00.022) 0:00:32.950 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.884) 0:00:33.834 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.026) 0:00:33.861 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.027) 0:00:33.888 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.054) 0:00:33.943 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.026) 0:00:33.969 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.026) 0:00:33.996 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.027) 0:00:34.023 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.025) 0:00:34.049 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.025) 0:00:34.075 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.053) 0:00:34.128 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.054) 0:00:34.183 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.038) 0:00:34.221 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.038) 0:00:34.260 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.049) 0:00:34.310 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.036) 0:00:34.346 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.037) 0:00:34.384 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.040) 0:00:34.424 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 08:12:30 +0000 (0:00:00.052) 0:00:34.477 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 08:12:31 +0000 (0:00:00.040) 0:00:34.518 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 08:12:31 +0000 (0:00:00.043) 0:00:34.562 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 08:12:31 +0000 (0:00:00.038) 0:00:34.600 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 08:12:31 +0000 (0:00:00.040) 0:00:34.641 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 08:12:31 +0000 (0:00:00.493) 0:00:35.135 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.417) 0:00:35.552 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.084) 0:00:35.637 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.076) 0:00:35.713 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.039) 0:00:35.753 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.076) 0:00:35.829 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.036) 0:00:35.866 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.037) 0:00:35.904 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.036) 0:00:35.940 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.036) 0:00:35.977 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.041) 0:00:36.018 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 08:12:32 +0000 (0:00:00.060) 0:00:36.078 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv1" ], "delta": "0:00:00.032714", "end": "2022-07-21 04:12:33.242528", "rc": 0, "start": "2022-07-21 04:12:33.209814" } STDOUT: LVM2_LV_NAME=lv1 LVM2_LV_ATTR=Vwi-aotz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.416) 0:00:36.495 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.053) 0:00:36.549 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.063) 0:00:36.613 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.043) 0:00:36.656 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.046) 0:00:36.702 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.041) 0:00:36.744 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:16 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.040) 0:00:36.784 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:43 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.040) 0:00:36.825 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:53 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.024) 0:00:36.849 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Repeat previous invocation to verify idempotence] ************************ task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:40 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.037) 0:00:36.887 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.052) 0:00:36.939 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:12:33 +0000 (0:00:00.037) 0:00:36.977 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:12:34 +0000 (0:00:00.515) 0:00:37.492 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:12:34 +0000 (0:00:00.066) 0:00:37.559 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:12:34 +0000 (0:00:00.073) 0:00:37.633 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:12:34 +0000 (0:00:00.100) 0:00:37.733 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:12:34 +0000 (0:00:00.047) 0:00:37.780 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:12:34 +0000 (0:00:00.022) 0:00:37.803 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:12:35 +0000 (0:00:00.841) 0:00:38.644 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "3g", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:12:35 +0000 (0:00:00.036) 0:00:38.681 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:12:35 +0000 (0:00:00.035) 0:00:38.716 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:12:37 +0000 (0:00:02.591) 0:00:41.308 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:12:37 +0000 (0:00:00.046) 0:00:41.355 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:12:37 +0000 (0:00:00.036) 0:00:41.392 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:12:37 +0000 (0:00:00.039) 0:00:41.431 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:12:37 +0000 (0:00:00.034) 0:00:41.465 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:12:38 +0000 (0:00:00.882) 0:00:42.348 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:12:40 +0000 (0:00:01.727) 0:00:44.075 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:12:40 +0000 (0:00:00.089) 0:00:44.165 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:12:40 +0000 (0:00:00.022) 0:00:44.188 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "xfsprogs", "e2fsprogs", "lvm2", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:12:43 +0000 (0:00:02.796) 0:00:46.984 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:12:43 +0000 (0:00:00.053) 0:00:47.038 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:12:43 +0000 (0:00:00.026) 0:00:47.064 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "xfsprogs", "e2fsprogs", "lvm2", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:12:43 +0000 (0:00:00.043) 0:00:47.108 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:12:43 +0000 (0:00:00.044) 0:00:47.152 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:12:43 +0000 (0:00:00.038) 0:00:47.190 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:12:43 +0000 (0:00:00.040) 0:00:47.231 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:12:44 +0000 (0:00:00.698) 0:00:47.930 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv1" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:12:44 +0000 (0:00:00.395) 0:00:48.325 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:12:45 +0000 (0:00:00.680) 0:00:49.006 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:12:45 +0000 (0:00:00.377) 0:00:49.383 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:12:45 +0000 (0:00:00.022) 0:00:49.406 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:56 Thursday 21 July 2022 08:12:46 +0000 (0:00:01.038) 0:00:50.445 ********* included: /tmp/tmpsvhdx31t/tests/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:1 Thursday 21 July 2022 08:12:47 +0000 (0:00:00.046) 0:00:50.491 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:6 Thursday 21 July 2022 08:12:47 +0000 (0:00:00.094) 0:00:50.585 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:14 Thursday 21 July 2022 08:12:47 +0000 (0:00:00.071) 0:00:50.657 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "jHWuRa-IE9u-VNaw-TG3b-eUEu-7c97-hAZTqb" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "CvMQQK-EGRf-g62R-VzcS-QyRJ-JQlf-3f0oiq" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "39QUtr-6fIa-Np89-vDzq-yFkf-XlPB-dklI1i" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-08-11-44-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:19 Thursday 21 July 2022 08:12:47 +0000 (0:00:00.377) 0:00:51.035 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002511", "end": "2022-07-21 04:12:48.149600", "rc": 0, "start": "2022-07-21 04:12:48.147089" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/vg1-lv1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 Thursday 21 July 2022 08:12:47 +0000 (0:00:00.364) 0:00:51.399 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002700", "end": "2022-07-21 04:12:48.508688", "failed_when_result": false, "rc": 1, "start": "2022-07-21 04:12:48.505988" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:33 Thursday 21 July 2022 08:12:48 +0000 (0:00:00.357) 0:00:51.757 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:5 Thursday 21 July 2022 08:12:48 +0000 (0:00:00.061) 0:00:51.818 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:18 Thursday 21 July 2022 08:12:48 +0000 (0:00:00.036) 0:00:51.855 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 08:12:48 +0000 (0:00:00.046) 0:00:51.902 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 08:12:48 +0000 (0:00:00.054) 0:00:51.956 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 08:12:49 +0000 (0:00:01.053) 0:00:53.010 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.062) 0:00:53.072 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.048) 0:00:53.121 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.084) 0:00:53.205 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.040) 0:00:53.245 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.052) 0:00:53.298 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.026) 0:00:53.325 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 08:12:49 +0000 (0:00:00.139) 0:00:53.464 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:6 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.044) 0:00:53.508 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:12 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.024) 0:00:53.533 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:16 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.029) 0:00:53.563 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:20 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.025) 0:00:53.588 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:24 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.025) 0:00:53.614 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:30 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.025) 0:00:53.639 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:36 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.026) 0:00:53.666 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:44 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.025) 0:00:53.691 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.036) 0:00:53.728 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.047) 0:00:53.775 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.045) 0:00:53.821 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.031) 0:00:53.852 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.029) 0:00:53.881 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.028) 0:00:53.910 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.047) 0:00:53.958 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.045) 0:00:54.003 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv1&&segtype=thin", "vg1" ], "delta": "0:00:00.035549", "end": "2022-07-21 04:12:51.155741", "rc": 0, "start": "2022-07-21 04:12:51.120192" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.401) 0:00:54.405 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 08:12:50 +0000 (0:00:00.064) 0:00:54.469 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.056) 0:00:54.525 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.042) 0:00:54.568 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.048) 0:00:54.617 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.054) 0:00:54.672 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.045) 0:00:54.718 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.053) 0:00:54.772 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.048) 0:00:54.820 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.088) 0:00:54.908 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.038) 0:00:54.947 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.082) 0:00:55.029 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.036) 0:00:55.066 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.037) 0:00:55.103 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.049) 0:00:55.153 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.054) 0:00:55.207 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.040) 0:00:55.248 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.039) 0:00:55.287 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.039) 0:00:55.326 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.040) 0:00:55.366 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.049) 0:00:55.416 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:12:51 +0000 (0:00:00.051) 0:00:55.468 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.040) 0:00:55.508 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.040) 0:00:55.548 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.039) 0:00:55.588 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.040) 0:00:55.629 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.038) 0:00:55.668 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.049) 0:00:55.717 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.049) 0:00:55.766 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.026) 0:00:55.793 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.025) 0:00:55.818 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.025) 0:00:55.844 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.026) 0:00:55.870 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.026) 0:00:55.896 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.026) 0:00:55.922 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.023) 0:00:55.946 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.038) 0:00:55.985 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.039) 0:00:56.024 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:2 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.047) 0:00:56.071 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:10 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.050) 0:00:56.122 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.074) 0:00:56.197 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.103) 0:00:56.300 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.058) 0:00:56.359 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.054) 0:00:56.414 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 08:12:52 +0000 (0:00:00.054) 0:00:56.468 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.051) 0:00:56.520 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.026) 0:00:56.547 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.028) 0:00:56.575 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.027) 0:00:56.603 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.039) 0:00:56.642 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/vg1-lv1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.067) 0:00:56.710 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.050) 0:00:56.760 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.052) 0:00:56.813 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.040) 0:00:56.854 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.036) 0:00:56.890 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.040) 0:00:56.931 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.043) 0:00:56.974 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658391137.2726676, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658391137.2726676, "dev": 5, "device_type": 64772, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1047, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658391137.2726676, "nlink": 1, "path": "/dev/mapper/vg1-lv1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.391) 0:00:57.366 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.039) 0:00:57.405 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 08:12:53 +0000 (0:00:00.042) 0:00:57.448 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 08:12:54 +0000 (0:00:00.036) 0:00:57.484 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 08:12:54 +0000 (0:00:00.024) 0:00:57.509 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 08:12:54 +0000 (0:00:00.039) 0:00:57.548 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 08:12:54 +0000 (0:00:00.026) 0:00:57.575 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 08:12:54 +0000 (0:00:00.888) 0:00:58.463 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.025) 0:00:58.489 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.024) 0:00:58.514 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.059) 0:00:58.574 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.038) 0:00:58.612 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.024) 0:00:58.637 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.024) 0:00:58.661 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.037) 0:00:58.699 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.033) 0:00:58.732 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.050) 0:00:58.783 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.049) 0:00:58.832 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.038) 0:00:58.871 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.037) 0:00:58.908 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.036) 0:00:58.945 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.038) 0:00:58.983 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.037) 0:00:59.021 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.037) 0:00:59.059 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.041) 0:00:59.100 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.038) 0:00:59.139 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.039) 0:00:59.178 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.040) 0:00:59.218 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 08:12:55 +0000 (0:00:00.039) 0:00:59.258 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.367) 0:00:59.625 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.369) 0:00:59.995 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.051) 0:01:00.046 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.039) 0:01:00.085 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.040) 0:01:00.126 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.039) 0:01:00.166 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.037) 0:01:00.203 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.037) 0:01:00.240 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.075) 0:01:00.316 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.039) 0:01:00.355 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.038) 0:01:00.393 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 08:12:56 +0000 (0:00:00.056) 0:01:00.450 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv1" ], "delta": "0:00:00.033870", "end": "2022-07-21 04:12:57.603369", "rc": 0, "start": "2022-07-21 04:12:57.569499" } STDOUT: LVM2_LV_NAME=lv1 LVM2_LV_ATTR=Vwi-aotz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.406) 0:01:00.856 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.052) 0:01:00.909 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.054) 0:01:00.963 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.040) 0:01:01.004 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.039) 0:01:01.044 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.037) 0:01:01.081 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:16 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.036) 0:01:01.118 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:43 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.038) 0:01:01.156 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:53 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.021) 0:01:01.178 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change thinlv fs type] *************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:58 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.035) 0:01:01.214 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.069) 0:01:01.283 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:12:57 +0000 (0:00:00.039) 0:01:01.322 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:12:58 +0000 (0:00:00.531) 0:01:01.853 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:12:58 +0000 (0:00:00.063) 0:01:01.917 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:12:58 +0000 (0:00:00.036) 0:01:01.954 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:12:58 +0000 (0:00:00.038) 0:01:01.992 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:12:58 +0000 (0:00:00.048) 0:01:02.041 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:12:58 +0000 (0:00:00.021) 0:01:02.062 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:12:59 +0000 (0:00:00.891) 0:01:02.954 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "fs_type": "xfs", "name": "lv1", "thin": true, "thin_pool_name": "tpool1" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:12:59 +0000 (0:00:00.050) 0:01:03.004 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:12:59 +0000 (0:00:00.036) 0:01:03.041 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2", "xfsprogs" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:13:02 +0000 (0:00:02.584) 0:01:05.625 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:13:02 +0000 (0:00:00.049) 0:01:05.674 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:13:02 +0000 (0:00:00.035) 0:01:05.710 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:13:02 +0000 (0:00:00.039) 0:01:05.750 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:13:02 +0000 (0:00:00.034) 0:01:05.785 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:13:03 +0000 (0:00:00.877) 0:01:06.662 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:13:04 +0000 (0:00:01.713) 0:01:08.376 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:13:04 +0000 (0:00:00.092) 0:01:08.469 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:13:05 +0000 (0:00:00.056) 0:01:08.525 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "path": "/opt/test1", "state": "absent" } ], "packages": [ "lvm2", "xfsprogs", "e2fsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:13:07 +0000 (0:00:02.749) 0:01:11.275 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:13:07 +0000 (0:00:00.036) 0:01:11.311 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:13:07 +0000 (0:00:00.022) 0:01:11.334 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "path": "/opt/test1", "state": "absent" } ], "packages": [ "lvm2", "xfsprogs", "e2fsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:13:07 +0000 (0:00:00.037) 0:01:11.372 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:13:07 +0000 (0:00:00.037) 0:01:11.409 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:13:07 +0000 (0:00:00.037) 0:01:11.447 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'path': '/opt/test1', 'state': 'absent'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": { "path": "/opt/test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:13:08 +0000 (0:00:00.401) 0:01:11.848 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:13:09 +0000 (0:00:00.697) 0:01:12.546 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:13:09 +0000 (0:00:00.039) 0:01:12.585 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:13:09 +0000 (0:00:00.667) 0:01:13.252 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:13:10 +0000 (0:00:00.362) 0:01:13.615 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:13:10 +0000 (0:00:00.023) 0:01:13.638 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:72 Thursday 21 July 2022 08:13:11 +0000 (0:00:01.052) 0:01:14.691 ********* included: /tmp/tmpsvhdx31t/tests/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:1 Thursday 21 July 2022 08:13:11 +0000 (0:00:00.046) 0:01:14.737 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:6 Thursday 21 July 2022 08:13:11 +0000 (0:00:00.051) 0:01:14.789 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:14 Thursday 21 July 2022 08:13:11 +0000 (0:00:00.036) 0:01:14.825 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "jHWuRa-IE9u-VNaw-TG3b-eUEu-7c97-hAZTqb" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "CvMQQK-EGRf-g62R-VzcS-QyRJ-JQlf-3f0oiq" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "39QUtr-6fIa-Np89-vDzq-yFkf-XlPB-dklI1i" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-08-11-44-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:19 Thursday 21 July 2022 08:13:11 +0000 (0:00:00.376) 0:01:15.202 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002495", "end": "2022-07-21 04:13:12.306117", "rc": 0, "start": "2022-07-21 04:13:12.303622" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 Thursday 21 July 2022 08:13:12 +0000 (0:00:00.352) 0:01:15.555 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.003843", "end": "2022-07-21 04:13:13.663318", "failed_when_result": false, "rc": 1, "start": "2022-07-21 04:13:12.659475" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:33 Thursday 21 July 2022 08:13:13 +0000 (0:00:01.365) 0:01:16.920 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:5 Thursday 21 July 2022 08:13:13 +0000 (0:00:00.063) 0:01:16.984 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:18 Thursday 21 July 2022 08:13:13 +0000 (0:00:00.035) 0:01:17.019 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 08:13:13 +0000 (0:00:00.050) 0:01:17.069 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 08:13:13 +0000 (0:00:00.100) 0:01:17.170 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 08:13:14 +0000 (0:00:01.043) 0:01:18.213 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 08:13:14 +0000 (0:00:00.085) 0:01:18.299 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 08:13:14 +0000 (0:00:00.051) 0:01:18.351 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 08:13:14 +0000 (0:00:00.085) 0:01:18.436 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.080) 0:01:18.516 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.086) 0:01:18.603 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.025) 0:01:18.628 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.155) 0:01:18.784 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:6 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.042) 0:01:18.826 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:12 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.025) 0:01:18.852 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:16 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.026) 0:01:18.879 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:20 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.025) 0:01:18.904 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:24 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.024) 0:01:18.929 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:30 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.026) 0:01:18.955 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:36 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.023) 0:01:18.979 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:44 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.023) 0:01:19.003 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.037) 0:01:19.041 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.044) 0:01:19.086 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.044) 0:01:19.130 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.031) 0:01:19.162 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.031) 0:01:19.193 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.030) 0:01:19.224 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.049) 0:01:19.273 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 08:13:15 +0000 (0:00:00.047) 0:01:19.321 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv1&&segtype=thin", "vg1" ], "delta": "0:00:00.041851", "end": "2022-07-21 04:13:16.479644", "rc": 0, "start": "2022-07-21 04:13:16.437793" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.410) 0:01:19.732 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.058) 0:01:19.791 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.053) 0:01:19.845 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.039) 0:01:19.885 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.048) 0:01:19.933 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.050) 0:01:19.984 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.032) 0:01:20.017 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.053) 0:01:20.071 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.088) 0:01:20.160 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.116) 0:01:20.277 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.037) 0:01:20.314 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.036) 0:01:20.350 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.036) 0:01:20.387 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.033) 0:01:20.421 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:13:16 +0000 (0:00:00.049) 0:01:20.470 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.048) 0:01:20.519 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.041) 0:01:20.560 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.039) 0:01:20.599 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.039) 0:01:20.639 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.039) 0:01:20.679 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.051) 0:01:20.731 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.053) 0:01:20.784 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.043) 0:01:20.827 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.039) 0:01:20.866 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.038) 0:01:20.905 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.039) 0:01:20.945 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.036) 0:01:20.982 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.051) 0:01:21.033 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.048) 0:01:21.082 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.027) 0:01:21.109 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.024) 0:01:21.134 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.024) 0:01:21.159 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.024) 0:01:21.183 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.025) 0:01:21.209 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.024) 0:01:21.233 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.024) 0:01:21.257 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.034) 0:01:21.292 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.036) 0:01:21.329 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:2 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.043) 0:01:21.372 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:10 Thursday 21 July 2022 08:13:17 +0000 (0:00:00.048) 0:01:21.421 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.075) 0:01:21.497 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.120) 0:01:21.617 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.058) 0:01:21.676 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.026) 0:01:21.703 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.054) 0:01:21.758 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.040) 0:01:21.798 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.026) 0:01:21.825 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.027) 0:01:21.853 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.026) 0:01:21.879 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.038) 0:01:21.917 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.065) 0:01:21.983 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.052) 0:01:22.035 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.053) 0:01:22.089 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.041) 0:01:22.130 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.037) 0:01:22.168 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.041) 0:01:22.210 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 08:13:18 +0000 (0:00:00.042) 0:01:22.253 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658391137.2726676, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658391137.2726676, "dev": 5, "device_type": 64772, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1047, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658391137.2726676, "nlink": 1, "path": "/dev/mapper/vg1-lv1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.372) 0:01:22.625 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.041) 0:01:22.667 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.043) 0:01:22.710 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.039) 0:01:22.749 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.026) 0:01:22.776 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.044) 0:01:22.821 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 08:13:19 +0000 (0:00:00.026) 0:01:22.847 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.903) 0:01:23.751 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.032) 0:01:23.783 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.037) 0:01:23.821 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.069) 0:01:23.890 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.034) 0:01:23.925 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.030) 0:01:23.956 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.025) 0:01:23.981 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.024) 0:01:24.005 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.024) 0:01:24.030 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.050) 0:01:24.081 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.050) 0:01:24.132 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.039) 0:01:24.171 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.035) 0:01:24.207 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.035) 0:01:24.243 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.035) 0:01:24.279 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.036) 0:01:24.316 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.036) 0:01:24.352 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.037) 0:01:24.389 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.040) 0:01:24.430 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 08:13:20 +0000 (0:00:00.037) 0:01:24.467 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.039) 0:01:24.507 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.036) 0:01:24.543 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.364) 0:01:24.907 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.363) 0:01:25.270 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.050) 0:01:25.321 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.036) 0:01:25.357 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.038) 0:01:25.396 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.037) 0:01:25.433 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 08:13:21 +0000 (0:00:00.037) 0:01:25.470 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.038) 0:01:25.508 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.077) 0:01:25.586 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.039) 0:01:25.625 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.041) 0:01:25.666 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.054) 0:01:25.721 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv1" ], "delta": "0:00:00.033339", "end": "2022-07-21 04:13:22.860490", "rc": 0, "start": "2022-07-21 04:13:22.827151" } STDOUT: LVM2_LV_NAME=lv1 LVM2_LV_ATTR=Vwi-a-tz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.390) 0:01:26.111 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.051) 0:01:26.163 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.050) 0:01:26.214 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.039) 0:01:26.254 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.039) 0:01:26.293 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.037) 0:01:26.331 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:16 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.037) 0:01:26.369 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:43 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.037) 0:01:26.406 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:53 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.023) 0:01:26.430 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create new LV under existing thinpool] *********************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:74 Thursday 21 July 2022 08:13:22 +0000 (0:00:00.045) 0:01:26.475 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.062) 0:01:26.537 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.037) 0:01:26.575 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.516) 0:01:27.091 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.066) 0:01:27.158 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.036) 0:01:27.194 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.037) 0:01:27.232 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.049) 0:01:27.282 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:13:23 +0000 (0:00:00.023) 0:01:27.305 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:13:24 +0000 (0:00:00.866) 0:01:28.172 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "mount_point": "/opt/test2", "name": "lv2", "size": "4g", "thin": true, "thin_pool_name": "tpool1" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:13:24 +0000 (0:00:00.040) 0:01:28.212 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:13:24 +0000 (0:00:00.035) 0:01:28.248 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:13:27 +0000 (0:00:02.569) 0:01:30.818 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:13:27 +0000 (0:00:00.050) 0:01:30.869 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:13:27 +0000 (0:00:00.037) 0:01:30.906 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:13:27 +0000 (0:00:00.043) 0:01:30.950 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:13:27 +0000 (0:00:00.038) 0:01:30.988 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:13:28 +0000 (0:00:00.842) 0:01:31.831 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:13:30 +0000 (0:00:01.658) 0:01:33.490 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:13:30 +0000 (0:00:00.101) 0:01:33.591 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:13:30 +0000 (0:00:00.024) 0:01:33.616 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "create device", "device": "/dev/mapper/vg1-lv2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "dosfstools", "lvm2" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:13:33 +0000 (0:00:02.997) 0:01:36.614 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:13:33 +0000 (0:00:00.040) 0:01:36.654 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:13:33 +0000 (0:00:00.023) 0:01:36.678 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "create device", "device": "/dev/mapper/vg1-lv2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "mounted" } ], "packages": [ "e2fsprogs", "xfsprogs", "dosfstools", "lvm2" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:13:33 +0000 (0:00:00.042) 0:01:36.720 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:13:33 +0000 (0:00:00.043) 0:01:36.763 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:13:33 +0000 (0:00:00.037) 0:01:36.801 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:13:33 +0000 (0:00:00.040) 0:01:36.842 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:13:34 +0000 (0:00:00.676) 0:01:37.519 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv2', 'path': '/opt/test2', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:13:34 +0000 (0:00:00.424) 0:01:37.943 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:13:35 +0000 (0:00:00.663) 0:01:38.607 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:13:35 +0000 (0:00:00.354) 0:01:38.961 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:13:35 +0000 (0:00:00.022) 0:01:38.984 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:89 Thursday 21 July 2022 08:13:36 +0000 (0:00:01.086) 0:01:40.070 ********* included: /tmp/tmpsvhdx31t/tests/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:1 Thursday 21 July 2022 08:13:36 +0000 (0:00:00.084) 0:01:40.155 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:6 Thursday 21 July 2022 08:13:36 +0000 (0:00:00.052) 0:01:40.207 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:14 Thursday 21 July 2022 08:13:36 +0000 (0:00:00.039) 0:01:40.247 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" }, "/dev/mapper/vg1-lv2": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv2", "size": "4G", "type": "lvm", "uuid": "75e0a969-5f13-4031-b338-13ed9ac6879c" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "jHWuRa-IE9u-VNaw-TG3b-eUEu-7c97-hAZTqb" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "CvMQQK-EGRf-g62R-VzcS-QyRJ-JQlf-3f0oiq" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "39QUtr-6fIa-Np89-vDzq-yFkf-XlPB-dklI1i" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-08-11-44-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:19 Thursday 21 July 2022 08:13:37 +0000 (0:00:00.376) 0:01:40.623 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002532", "end": "2022-07-21 04:13:37.734485", "rc": 0, "start": "2022-07-21 04:13:37.731953" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/vg1-lv2 /opt/test2 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 Thursday 21 July 2022 08:13:37 +0000 (0:00:00.362) 0:01:40.985 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002524", "end": "2022-07-21 04:13:38.091539", "failed_when_result": false, "rc": 1, "start": "2022-07-21 04:13:38.089015" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:33 Thursday 21 July 2022 08:13:37 +0000 (0:00:00.358) 0:01:41.344 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:5 Thursday 21 July 2022 08:13:37 +0000 (0:00:00.061) 0:01:41.406 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:18 Thursday 21 July 2022 08:13:37 +0000 (0:00:00.037) 0:01:41.443 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 08:13:38 +0000 (0:00:00.050) 0:01:41.494 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 08:13:38 +0000 (0:00:00.058) 0:01:41.552 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 08:13:39 +0000 (0:00:01.045) 0:01:42.598 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.092) 0:01:42.691 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.053) 0:01:42.744 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.083) 0:01:42.827 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.079) 0:01:42.907 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.088) 0:01:42.996 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.028) 0:01:43.024 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.102) 0:01:43.127 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:6 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.045) 0:01:43.173 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:12 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.027) 0:01:43.200 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:16 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.027) 0:01:43.227 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:20 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.028) 0:01:43.256 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:24 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.026) 0:01:43.283 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:30 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.027) 0:01:43.310 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:36 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.028) 0:01:43.339 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:44 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.026) 0:01:43.365 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.034) 0:01:43.400 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 08:13:39 +0000 (0:00:00.046) 0:01:43.447 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.045) 0:01:43.492 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.032) 0:01:43.525 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.029) 0:01:43.555 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.027) 0:01:43.582 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.046) 0:01:43.629 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.046) 0:01:43.676 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv2&&segtype=thin", "vg1" ], "delta": "0:00:00.038215", "end": "2022-07-21 04:13:40.840447", "rc": 0, "start": "2022-07-21 04:13:40.802232" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.418) 0:01:44.094 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.058) 0:01:44.153 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.058) 0:01:44.212 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.040) 0:01:44.252 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.048) 0:01:44.301 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.083) 0:01:44.385 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 08:13:40 +0000 (0:00:00.079) 0:01:44.464 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.053) 0:01:44.517 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.062) 0:01:44.579 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.051) 0:01:44.631 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.038) 0:01:44.669 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.035) 0:01:44.704 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.036) 0:01:44.741 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.036) 0:01:44.777 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.051) 0:01:44.828 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.051) 0:01:44.880 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.036) 0:01:44.916 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.036) 0:01:44.953 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.039) 0:01:44.993 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.037) 0:01:45.030 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.053) 0:01:45.083 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.054) 0:01:45.137 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.039) 0:01:45.177 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.038) 0:01:45.216 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.040) 0:01:45.256 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.037) 0:01:45.294 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.037) 0:01:45.331 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.052) 0:01:45.384 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.049) 0:01:45.433 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 08:13:41 +0000 (0:00:00.026) 0:01:45.460 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.026) 0:01:45.487 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.028) 0:01:45.515 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.028) 0:01:45.544 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.027) 0:01:45.571 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.028) 0:01:45.599 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.060) 0:01:45.660 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.038) 0:01:45.698 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.041) 0:01:45.739 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:2 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.053) 0:01:45.793 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:10 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.056) 0:01:45.850 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.077) 0:01:45.927 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.045) 0:01:45.973 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1030395, "block_size": 4096, "block_total": 1046016, "block_used": 15621, "device": "/dev/mapper/vg1-lv2", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 4220497920, "size_total": 4284481536, "uuid": "75e0a969-5f13-4031-b338-13ed9ac6879c" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1030395, "block_size": 4096, "block_total": 1046016, "block_used": 15621, "device": "/dev/mapper/vg1-lv2", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 4220497920, "size_total": 4284481536, "uuid": "75e0a969-5f13-4031-b338-13ed9ac6879c" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.059) 0:01:46.032 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.053) 0:01:46.086 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.049) 0:01:46.136 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.056) 0:01:46.192 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.025) 0:01:46.218 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.024) 0:01:46.243 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.025) 0:01:46.268 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.034) 0:01:46.302 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/vg1-lv2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.062) 0:01:46.365 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.052) 0:01:46.417 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 08:13:42 +0000 (0:00:00.048) 0:01:46.466 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.036) 0:01:46.502 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.036) 0:01:46.539 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.038) 0:01:46.577 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.040) 0:01:46.618 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658391213.2956676, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658391213.2956676, "dev": 5, "device_type": 64773, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1629, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658391213.2956676, "nlink": 1, "path": "/dev/mapper/vg1-lv2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.366) 0:01:46.985 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.041) 0:01:47.026 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.082) 0:01:47.109 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.107) 0:01:47.216 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.026) 0:01:47.243 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.040) 0:01:47.284 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 08:13:43 +0000 (0:00:00.025) 0:01:47.309 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.848) 0:01:48.158 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.026) 0:01:48.184 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.024) 0:01:48.209 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.052) 0:01:48.261 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.024) 0:01:48.286 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.025) 0:01:48.312 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.025) 0:01:48.338 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.026) 0:01:48.364 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.028) 0:01:48.393 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 08:13:44 +0000 (0:00:00.053) 0:01:48.446 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.051) 0:01:48.498 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.042) 0:01:48.541 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.040) 0:01:48.581 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.041) 0:01:48.622 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.040) 0:01:48.663 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.041) 0:01:48.704 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.039) 0:01:48.744 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.040) 0:01:48.785 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.040) 0:01:48.826 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.041) 0:01:48.868 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.040) 0:01:48.909 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.093) 0:01:49.002 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 08:13:45 +0000 (0:00:00.353) 0:01:49.355 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.369) 0:01:49.725 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.055) 0:01:49.780 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.039) 0:01:49.820 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.041) 0:01:49.862 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.040) 0:01:49.903 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.040) 0:01:49.943 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.044) 0:01:49.988 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.044) 0:01:50.032 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.044) 0:01:50.076 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.044) 0:01:50.121 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 08:13:46 +0000 (0:00:00.053) 0:01:50.174 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv2" ], "delta": "0:00:00.033579", "end": "2022-07-21 04:13:47.312587", "rc": 0, "start": "2022-07-21 04:13:47.279008" } STDOUT: LVM2_LV_NAME=lv2 LVM2_LV_ATTR=Vwi-aotz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.391) 0:01:50.565 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.051) 0:01:50.617 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.062) 0:01:50.679 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.040) 0:01:50.719 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.040) 0:01:50.760 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.044) 0:01:50.804 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:16 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.042) 0:01:50.847 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:43 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.036) 0:01:50.883 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:53 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.027) 0:01:50.911 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove existing LV under existing thinpool] ****************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:91 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.130) 0:01:51.041 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.068) 0:01:51.109 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:13:47 +0000 (0:00:00.039) 0:01:51.149 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:13:48 +0000 (0:00:00.506) 0:01:51.655 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:13:48 +0000 (0:00:00.068) 0:01:51.724 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:13:48 +0000 (0:00:00.040) 0:01:51.764 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:13:48 +0000 (0:00:00.039) 0:01:51.804 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:13:48 +0000 (0:00:00.050) 0:01:51.854 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:13:48 +0000 (0:00:00.022) 0:01:51.877 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:13:49 +0000 (0:00:00.833) 0:01:52.710 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "mount_point": "/opt/test2", "name": "lv2", "state": "absent", "thin": true, "thin_pool_name": "tpool1" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:13:49 +0000 (0:00:00.041) 0:01:52.752 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:13:49 +0000 (0:00:00.038) 0:01:52.790 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:13:51 +0000 (0:00:02.650) 0:01:55.440 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:13:52 +0000 (0:00:00.049) 0:01:55.490 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:13:52 +0000 (0:00:00.071) 0:01:55.562 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:13:52 +0000 (0:00:00.041) 0:01:55.603 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:13:52 +0000 (0:00:00.036) 0:01:55.640 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:13:52 +0000 (0:00:00.836) 0:01:56.476 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:13:54 +0000 (0:00:01.636) 0:01:58.112 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:13:54 +0000 (0:00:00.058) 0:01:58.171 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:13:54 +0000 (0:00:00.022) 0:01:58.194 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv2", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "absent" } ], "packages": [ "xfsprogs", "lvm2", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:13:58 +0000 (0:00:03.288) 0:02:01.483 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:13:58 +0000 (0:00:00.039) 0:02:01.522 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:13:58 +0000 (0:00:00.025) 0:02:01.547 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv2", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "absent" } ], "packages": [ "xfsprogs", "lvm2", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:13:58 +0000 (0:00:00.045) 0:02:01.593 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:13:58 +0000 (0:00:00.079) 0:02:01.672 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:13:58 +0000 (0:00:00.042) 0:02:01.715 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv2', 'path': '/opt/test2', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv2" } TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:13:58 +0000 (0:00:00.429) 0:02:02.144 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:13:59 +0000 (0:00:00.650) 0:02:02.795 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:13:59 +0000 (0:00:00.041) 0:02:02.836 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:14:00 +0000 (0:00:00.705) 0:02:03.541 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:14:00 +0000 (0:00:00.356) 0:02:03.898 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:14:00 +0000 (0:00:00.024) 0:02:03.922 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:106 Thursday 21 July 2022 08:14:01 +0000 (0:00:01.002) 0:02:04.925 ********* included: /tmp/tmpsvhdx31t/tests/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:1 Thursday 21 July 2022 08:14:01 +0000 (0:00:00.053) 0:02:04.978 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:6 Thursday 21 July 2022 08:14:01 +0000 (0:00:00.054) 0:02:05.032 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:14 Thursday 21 July 2022 08:14:01 +0000 (0:00:00.038) 0:02:05.071 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "fb6ea06c-ae8b-47ef-ac11-0cd3386d2000" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "jHWuRa-IE9u-VNaw-TG3b-eUEu-7c97-hAZTqb" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "CvMQQK-EGRf-g62R-VzcS-QyRJ-JQlf-3f0oiq" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "39QUtr-6fIa-Np89-vDzq-yFkf-XlPB-dklI1i" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-08-11-44-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:19 Thursday 21 July 2022 08:14:01 +0000 (0:00:00.372) 0:02:05.444 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.003014", "end": "2022-07-21 04:14:02.553978", "rc": 0, "start": "2022-07-21 04:14:02.550964" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 Thursday 21 July 2022 08:14:02 +0000 (0:00:00.367) 0:02:05.812 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.003573", "end": "2022-07-21 04:14:03.977080", "failed_when_result": false, "rc": 1, "start": "2022-07-21 04:14:02.973507" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:33 Thursday 21 July 2022 08:14:03 +0000 (0:00:01.418) 0:02:07.231 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:5 Thursday 21 July 2022 08:14:03 +0000 (0:00:00.061) 0:02:07.293 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:18 Thursday 21 July 2022 08:14:03 +0000 (0:00:00.036) 0:02:07.329 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 08:14:03 +0000 (0:00:00.078) 0:02:07.408 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 08:14:03 +0000 (0:00:00.058) 0:02:07.466 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 08:14:04 +0000 (0:00:01.000) 0:02:08.467 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.054) 0:02:08.521 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.064) 0:02:08.585 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.053) 0:02:08.639 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.042) 0:02:08.682 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.051) 0:02:08.733 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.026) 0:02:08.760 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.072) 0:02:08.833 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:6 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.043) 0:02:08.876 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:12 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.026) 0:02:08.902 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:16 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.026) 0:02:08.929 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:20 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.026) 0:02:08.956 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:24 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.024) 0:02:08.981 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:30 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.026) 0:02:09.007 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:36 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.024) 0:02:09.031 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:44 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.023) 0:02:09.055 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.035) 0:02:09.090 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.045) 0:02:09.136 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.043) 0:02:09.179 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.029) 0:02:09.208 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.031) 0:02:09.239 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.030) 0:02:09.270 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.078) 0:02:09.348 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.046) 0:02:09.395 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.028) 0:02:09.423 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 08:14:05 +0000 (0:00:00.028) 0:02:09.451 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.029) 0:02:09.480 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.027) 0:02:09.508 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.048) 0:02:09.557 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.052) 0:02:09.610 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.036) 0:02:09.647 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.054) 0:02:09.701 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.051) 0:02:09.753 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.052) 0:02:09.805 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.039) 0:02:09.844 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.037) 0:02:09.881 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.041) 0:02:09.923 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.037) 0:02:09.960 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.051) 0:02:10.011 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.049) 0:02:10.061 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.036) 0:02:10.098 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.036) 0:02:10.135 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.041) 0:02:10.176 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.037) 0:02:10.214 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:6 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.052) 0:02:10.266 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:11 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.051) 0:02:10.317 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:17 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.038) 0:02:10.356 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:23 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.038) 0:02:10.394 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-crypttab.yml:29 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.043) 0:02:10.437 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 08:14:06 +0000 (0:00:00.034) 0:02:10.472 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.035) 0:02:10.507 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.051) 0:02:10.559 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.079) 0:02:10.638 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.027) 0:02:10.665 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.026) 0:02:10.692 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.026) 0:02:10.719 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.027) 0:02:10.747 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.026) 0:02:10.773 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.026) 0:02:10.799 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.027) 0:02:10.827 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.042) 0:02:10.870 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.038) 0:02:10.908 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:2 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.046) 0:02:10.955 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:10 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.051) 0:02:11.007 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.076) 0:02:11.084 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.043) 0:02:11.128 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.059) 0:02:11.187 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.027) 0:02:11.214 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.051) 0:02:11.266 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.039) 0:02:11.305 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.027) 0:02:11.333 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.026) 0:02:11.359 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.026) 0:02:11.386 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 08:14:07 +0000 (0:00:00.039) 0:02:11.425 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.067) 0:02:11.493 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.025) 0:02:11.518 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.052) 0:02:11.571 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.043) 0:02:11.615 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.037) 0:02:11.652 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.026) 0:02:11.679 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.027) 0:02:11.707 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.353) 0:02:12.060 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.082) 0:02:12.143 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.026) 0:02:12.169 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.077) 0:02:12.246 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.026) 0:02:12.272 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.025) 0:02:12.298 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 08:14:08 +0000 (0:00:00.026) 0:02:12.325 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.850) 0:02:13.175 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.027) 0:02:13.202 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.026) 0:02:13.228 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.025) 0:02:13.254 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.028) 0:02:13.282 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.026) 0:02:13.309 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.027) 0:02:13.336 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.026) 0:02:13.362 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.028) 0:02:13.391 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 08:14:09 +0000 (0:00:00.053) 0:02:13.444 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.051) 0:02:13.495 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.038) 0:02:13.533 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.035) 0:02:13.569 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.035) 0:02:13.605 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.037) 0:02:13.642 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.037) 0:02:13.679 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.036) 0:02:13.716 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.037) 0:02:13.754 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.036) 0:02:13.791 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.041) 0:02:13.832 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.038) 0:02:13.871 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.038) 0:02:13.909 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.025) 0:02:13.935 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.037) 0:02:13.973 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.095) 0:02:14.069 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.036) 0:02:14.105 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.036) 0:02:14.141 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.041) 0:02:14.183 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.037) 0:02:14.220 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.037) 0:02:14.257 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.040) 0:02:14.298 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.035) 0:02:14.334 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.036) 0:02:14.370 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.025) 0:02:14.396 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.024) 0:02:14.421 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.024) 0:02:14.446 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 08:14:10 +0000 (0:00:00.024) 0:02:14.470 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.025) 0:02:14.496 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.024) 0:02:14.520 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.025) 0:02:14.545 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:16 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.025) 0:02:14.570 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:43 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.038) 0:02:14.609 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:53 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.022) 0:02:14.632 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Cleanup] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:108 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.035) 0:02:14.667 ********* TASK [linux-system-roles.storage : set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:2 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.077) 0:02:14.744 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : Ensure ansible_facts used by role] ********** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:2 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.039) 0:02:14.783 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [linux-system-roles.storage : Set platform/version specific variables] **** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/set_vars.yml:8 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.506) 0:02:15.290 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:5 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.068) 0:02:15.359 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [linux-system-roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:9 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.036) 0:02:15.395 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : include the appropriate provider tasks] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main.yml:13 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.033) 0:02:15.429 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 08:14:11 +0000 (0:00:00.047) 0:02:15.476 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : make sure blivet is available] ************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 08:14:12 +0000 (0:00:00.062) 0:02:15.538 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : show storage_pools] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:14 Thursday 21 July 2022 08:14:12 +0000 (0:00:00.824) 0:02:16.363 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "state": "absent", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "3g", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g" } ] } ] } TASK [linux-system-roles.storage : show storage_volumes] *********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:19 Thursday 21 July 2022 08:14:12 +0000 (0:00:00.039) 0:02:16.403 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [linux-system-roles.storage : get required packages] ********************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 Thursday 21 July 2022 08:14:12 +0000 (0:00:00.037) 0:02:16.441 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [linux-system-roles.storage : enable copr repositories if needed] ********* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:37 Thursday 21 July 2022 08:14:15 +0000 (0:00:02.450) 0:02:18.891 ********* included: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [linux-system-roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 08:14:15 +0000 (0:00:00.046) 0:02:18.937 ********* TASK [linux-system-roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 08:14:15 +0000 (0:00:00.033) 0:02:18.971 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : enable COPRs] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 08:14:15 +0000 (0:00:00.037) 0:02:19.008 ********* TASK [linux-system-roles.storage : make sure required packages are installed] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:44 Thursday 21 July 2022 08:14:15 +0000 (0:00:00.036) 0:02:19.045 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [linux-system-roles.storage : get service facts] ************************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Thursday 21 July 2022 08:14:16 +0000 (0:00:00.796) 0:02:19.842 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [linux-system-roles.storage : Set storage_cryptsetup_services] ************ task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 08:14:17 +0000 (0:00:01.627) 0:02:21.469 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [linux-system-roles.storage : Mask the systemd cryptsetup services] ******* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:71 Thursday 21 July 2022 08:14:18 +0000 (0:00:00.087) 0:02:21.557 ********* TASK [linux-system-roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 Thursday 21 July 2022 08:14:18 +0000 (0:00:00.021) 0:02:21.579 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "destroy device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "destroy device", "device": "/dev/vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdc1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdb1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "disklabel" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [], "packages": [ "xfsprogs", "e2fsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [linux-system-roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:91 Thursday 21 July 2022 08:14:22 +0000 (0:00:04.219) 0:02:25.798 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : Unmask the systemd cryptsetup services] ***** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:103 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.037) 0:02:25.836 ********* TASK [linux-system-roles.storage : show blivet_output] ************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:109 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.022) 0:02:25.859 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "destroy device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "destroy device", "device": "/dev/vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdc1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdb1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "disklabel" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/sda", "/dev/sdb", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [], "packages": [ "xfsprogs", "e2fsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [linux-system-roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:114 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.038) 0:02:25.897 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [linux-system-roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:118 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.038) 0:02:25.935 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [linux-system-roles.storage : remove obsolete mounts] ********************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:134 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.037) 0:02:25.973 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:146 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.034) 0:02:26.008 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : set up new/current mounts] ****************** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:151 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.024) 0:02:26.032 ********* TASK [linux-system-roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:163 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.036) 0:02:26.068 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [linux-system-roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:171 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.023) 0:02:26.091 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [linux-system-roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:176 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.352) 0:02:26.443 ********* TASK [linux-system-roles.storage : Update facts] ******************************* task path: /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:198 Thursday 21 July 2022 08:14:22 +0000 (0:00:00.022) 0:02:26.466 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/tests_create_thinp_then_remove.yml:125 Thursday 21 July 2022 08:14:23 +0000 (0:00:00.963) 0:02:27.430 ********* included: /tmp/tmpsvhdx31t/tests/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:1 Thursday 21 July 2022 08:14:24 +0000 (0:00:00.053) 0:02:27.483 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:6 Thursday 21 July 2022 08:14:24 +0000 (0:00:00.084) 0:02:27.567 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:14 Thursday 21 July 2022 08:14:24 +0000 (0:00:00.035) 0:02:27.603 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-08-11-44-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:19 Thursday 21 July 2022 08:14:24 +0000 (0:00:00.401) 0:02:28.004 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002424", "end": "2022-07-21 04:14:25.153959", "rc": 0, "start": "2022-07-21 04:14:25.151535" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 Thursday 21 July 2022 08:14:24 +0000 (0:00:00.411) 0:02:28.415 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004203", "end": "2022-07-21 04:14:26.521471", "failed_when_result": false, "rc": 1, "start": "2022-07-21 04:14:25.517268" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:33 Thursday 21 July 2022 08:14:26 +0000 (0:00:01.360) 0:02:29.775 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'absent', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:5 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.060) 0:02:29.836 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool.yml:18 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.038) 0:02:29.875 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:1 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.046) 0:02:29.921 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:10 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.054) 0:02:29.976 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:19 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.023) 0:02:30.000 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:23 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.049) 0:02:30.049 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:27 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.052) 0:02:30.102 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:34 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.054) 0:02:30.157 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:38 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.040) 0:02:30.197 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:42 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.051) 0:02:30.249 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:46 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.026) 0:02:30.276 ********* TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:56 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.022) 0:02:30.298 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:6 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.041) 0:02:30.340 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:12 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.025) 0:02:30.366 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:16 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.025) 0:02:30.391 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:20 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.025) 0:02:30.416 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:24 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.025) 0:02:30.442 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:30 Thursday 21 July 2022 08:14:26 +0000 (0:00:00.026) 0:02:30.468 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:36 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.025) 0:02:30.494 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-md.yml:44 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.024) 0:02:30.519 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:59 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.071) 0:02:30.590 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.071) 0:02:30.662 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.044) 0:02:30.706 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.031) 0:02:30.737 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.028) 0:02:30.766 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:62 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.027) 0:02:30.793 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-thin.yml:1 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.044) 0:02:30.837 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:3 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.040) 0:02:30.878 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:8 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:30.901 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:13 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:30.925 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-thin.yml:17 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:30.948 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:65 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:30.971 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.045) 0:02:31.017 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.046) 0:02:31.063 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.020) 0:02:31.084 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.021) 0:02:31.105 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:68 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.031) 0:02:31.137 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.046) 0:02:31.184 ********* included: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.043) 0:02:31.227 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:31.250 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.022) 0:02:31.273 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:31.297 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.022) 0:02:31.319 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.023) 0:02:31.342 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.025) 0:02:31.368 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.027) 0:02:31.395 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-members.yml:71 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.039) 0:02:31.435 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 08:14:27 +0000 (0:00:00.037) 0:02:31.473 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sda', 'sdb', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:2 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.041) 0:02:31.514 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:10 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.046) 0:02:31.561 ********* included: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:6 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.072) 0:02:31.633 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:14 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.082) 0:02:31.716 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:28 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.129) 0:02:31.846 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:37 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.025) 0:02:31.871 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:45 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.052) 0:02:31.924 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:54 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.042) 0:02:31.967 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:58 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.025) 0:02:31.993 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:63 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.025) 0:02:32.018 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-mount.yml:75 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.025) 0:02:32.044 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.040) 0:02:32.084 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.063) 0:02:32.147 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:32 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.025) 0:02:32.173 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:39 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.054) 0:02:32.228 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fstab.yml:49 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.038) 0:02:32.267 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:4 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.037) 0:02:32.304 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-fs.yml:10 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.025) 0:02:32.330 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:4 Thursday 21 July 2022 08:14:28 +0000 (0:00:00.026) 0:02:32.357 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:10 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.345) 0:02:32.702 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:18 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.042) 0:02:32.745 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:24 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.024) 0:02:32.770 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:28 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.040) 0:02:32.811 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-device.yml:33 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.024) 0:02:32.835 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.022) 0:02:32.858 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 08:14:29 +0000 (0:00:00.023) 0:02:32.881 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.842) 0:02:33.724 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.026) 0:02:33.750 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:30 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.027) 0:02:33.777 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:38 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.024) 0:02:33.802 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.025) 0:02:33.828 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:49 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.026) 0:02:33.855 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:55 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.027) 0:02:33.882 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:61 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.025) 0:02:33.907 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.025) 0:02:33.932 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:74 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.088) 0:02:34.021 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:79 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.088) 0:02:34.109 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:85 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.039) 0:02:34.148 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:91 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.039) 0:02:34.188 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-encryption.yml:97 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.076) 0:02:34.265 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:7 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.036) 0:02:34.302 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:13 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.039) 0:02:34.342 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:17 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.040) 0:02:34.382 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:21 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.040) 0:02:34.422 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:25 Thursday 21 July 2022 08:14:30 +0000 (0:00:00.038) 0:02:34.461 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:31 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.039) 0:02:34.500 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-md.yml:37 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.040) 0:02:34.540 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:3 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.037) 0:02:34.578 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:9 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.025) 0:02:34.604 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:15 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.037) 0:02:34.641 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:20 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.039) 0:02:34.680 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:25 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.036) 0:02:34.717 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:28 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.037) 0:02:34.754 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:31 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.037) 0:02:34.791 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:36 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.039) 0:02:34.830 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:39 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.037) 0:02:34.868 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:44 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.036) 0:02:34.905 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:47 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.037) 0:02:34.942 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-size.yml:50 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.042) 0:02:34.985 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:6 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.025) 0:02:35.010 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:14 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.023) 0:02:35.034 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:17 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.024) 0:02:35.058 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:22 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.025) 0:02:35.084 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:26 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.023) 0:02:35.108 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:32 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.023) 0:02:35.131 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume-cache.yml:36 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.023) 0:02:35.155 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpsvhdx31t/tests/test-verify-volume.yml:16 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.024) 0:02:35.179 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:43 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.033) 0:02:35.213 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpsvhdx31t/tests/verify-role-results.yml:53 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.021) 0:02:35.234 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-9.qcow2.snap : ok=646 changed=8 unreachable=0 failed=0 skipped=476 rescued=0 ignored=0 Thursday 21 July 2022 08:14:31 +0000 (0:00:00.047) 0:02:35.281 ********* =============================================================================== linux-system-roles.storage : manage the pools and volumes to match the specified state --- 4.54s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 4.22s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 3.29s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 3.00s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.80s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : manage the pools and volumes to match the specified state --- 2.75s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:77 linux-system-roles.storage : get required packages ---------------------- 2.65s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 2.59s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 2.58s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 2.57s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get required packages ---------------------- 2.45s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 linux-system-roles.storage : get service facts -------------------------- 1.80s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.75s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.73s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.71s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.66s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.64s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 linux-system-roles.storage : get service facts -------------------------- 1.63s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:51 Read the /etc/crypttab file --------------------------------------------- 1.42s /tmp/tmpsvhdx31t/tests/verify-role-results.yml:24 ----------------------------- linux-system-roles.storage : get required packages ---------------------- 1.37s /tmp/tmpsvhdx31t/tests/roles/linux-system-roles.storage/tasks/main-blivet.yml:24 ansible-playbook [core 2.12.6] config file = /etc/ansible/ansible.cfg configured module search path = ['/root/.ansible/plugins/modules', '/usr/share/ansible/plugins/modules'] ansible python module location = /usr/lib/python3.9/site-packages/ansible ansible collection location = /tmp/tmpi19f9hzy executable location = /usr/bin/ansible-playbook python version = 3.9.13 (main, May 18 2022, 00:00:00) [GCC 11.3.1 20220421 (Red Hat 11.3.1-2)] jinja version = 2.11.3 libyaml = True Using /etc/ansible/ansible.cfg as config file Skipping callback 'debug', as we already have a stdout callback. Skipping callback 'default', as we already have a stdout callback. Skipping callback 'minimal', as we already have a stdout callback. Skipping callback 'oneline', as we already have a stdout callback. PLAYBOOK: tests_create_thinp_then_remove.yml *********************************** 1 plays in /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml PLAY [all] ********************************************************************* TASK [Gathering Facts] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:2 Thursday 21 July 2022 11:24:37 +0000 (0:00:00.032) 0:00:00.032 ********* ok: [/cache/rhel-9.qcow2.snap] META: ran handlers TASK [include_role : fedora.linux_system_roles.storage] ************************ task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:14 Thursday 21 July 2022 11:24:38 +0000 (0:00:01.344) 0:00:01.376 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:24:38 +0000 (0:00:00.038) 0:00:01.415 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:24:38 +0000 (0:00:00.030) 0:00:01.445 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:24:39 +0000 (0:00:00.505) 0:00:01.951 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:24:39 +0000 (0:00:00.061) 0:00:02.012 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:24:39 +0000 (0:00:00.040) 0:00:02.053 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:24:39 +0000 (0:00:00.030) 0:00:02.083 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:24:39 +0000 (0:00:00.055) 0:00:02.139 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:24:39 +0000 (0:00:00.019) 0:00:02.158 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:24:40 +0000 (0:00:01.239) 0:00:03.398 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": "VARIABLE IS NOT DEFINED!: 'storage_pools' is undefined" } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:24:40 +0000 (0:00:00.033) 0:00:03.432 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:24:40 +0000 (0:00:00.031) 0:00:03.464 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:24:41 +0000 (0:00:00.701) 0:00:04.165 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:24:41 +0000 (0:00:00.042) 0:00:04.208 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:24:41 +0000 (0:00:00.029) 0:00:04.237 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:24:41 +0000 (0:00:00.033) 0:00:04.271 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:24:41 +0000 (0:00:00.028) 0:00:04.299 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:24:42 +0000 (0:00:00.846) 0:00:05.146 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:24:45 +0000 (0:00:02.828) 0:00:07.975 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.099) 0:00:08.074 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.048) 0:00:08.122 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.528) 0:00:08.651 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.036) 0:00:08.687 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.023) 0:00:08.710 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.039) 0:00:08.750 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.039) 0:00:08.789 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.036) 0:00:08.826 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.035) 0:00:08.861 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.022) 0:00:08.883 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:24:45 +0000 (0:00:00.034) 0:00:08.918 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:24:46 +0000 (0:00:00.021) 0:00:08.940 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:24:46 +0000 (0:00:00.492) 0:00:09.432 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:24:46 +0000 (0:00:00.021) 0:00:09.454 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:17 Thursday 21 July 2022 11:24:47 +0000 (0:00:00.972) 0:00:10.426 ********* included: /tmp/tmpfc26zqih/tests/storage/get_unused_disk.yml for /cache/rhel-9.qcow2.snap TASK [Find unused disks in the system] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/get_unused_disk.yml:2 Thursday 21 July 2022 11:24:47 +0000 (0:00:00.036) 0:00:10.462 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "disks": [ "sda", "sdb", "sdc" ] } TASK [Set unused_disks if necessary] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/get_unused_disk.yml:9 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.514) 0:00:10.977 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "unused_disks": [ "sda", "sdb", "sdc" ] }, "changed": false } TASK [Exit playbook when there's not enough unused disks in the system] ******** task path: /tmp/tmpfc26zqih/tests/storage/get_unused_disk.yml:14 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.036) 0:00:11.014 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Print unused disks] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/get_unused_disk.yml:19 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.039) 0:00:11.054 ********* ok: [/cache/rhel-9.qcow2.snap] => { "unused_disks": [ "sda", "sdb", "sdc" ] } TASK [Create a thinpool device] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:21 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.037) 0:00:11.091 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.045) 0:00:11.137 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.036) 0:00:11.173 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.513) 0:00:11.687 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.060) 0:00:11.747 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.033) 0:00:11.780 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.033) 0:00:11.814 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.053) 0:00:11.867 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:24:48 +0000 (0:00:00.021) 0:00:11.889 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:24:49 +0000 (0:00:00.870) 0:00:12.759 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "state": "present", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "3g", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:24:49 +0000 (0:00:00.039) 0:00:12.799 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:24:49 +0000 (0:00:00.038) 0:00:12.837 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:24:51 +0000 (0:00:01.327) 0:00:14.164 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:24:51 +0000 (0:00:00.052) 0:00:14.217 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:24:51 +0000 (0:00:00.067) 0:00:14.285 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:24:51 +0000 (0:00:00.040) 0:00:14.325 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:24:51 +0000 (0:00:00.034) 0:00:14.360 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:24:52 +0000 (0:00:00.829) 0:00:15.189 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "stopped", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:24:53 +0000 (0:00:01.629) 0:00:16.819 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:24:53 +0000 (0:00:00.055) 0:00:16.874 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:24:53 +0000 (0:00:00.022) 0:00:16.897 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "create format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdc1", "fs_type": null }, { "action": "create format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "dosfstools", "xfsprogs", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:24:58 +0000 (0:00:04.333) 0:00:21.231 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:24:58 +0000 (0:00:00.036) 0:00:21.267 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:24:58 +0000 (0:00:00.023) 0:00:21.291 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "create format", "device": "/dev/sdc", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdc1", "fs_type": null }, { "action": "create format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sdb1", "fs_type": null }, { "action": "create format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "create format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "create device", "device": "/dev/sda1", "fs_type": null }, { "action": "create format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "create device", "device": "/dev/vg1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "create device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv1" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "dosfstools", "xfsprogs", "lvm2", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:24:58 +0000 (0:00:00.041) 0:00:21.332 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:24:58 +0000 (0:00:00.037) 0:00:21.369 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:24:58 +0000 (0:00:00.039) 0:00:21.408 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:24:58 +0000 (0:00:00.041) 0:00:21.450 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:24:59 +0000 (0:00:00.977) 0:00:22.427 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:25:00 +0000 (0:00:00.551) 0:00:22.979 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:25:00 +0000 (0:00:00.708) 0:00:23.688 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:25:01 +0000 (0:00:00.412) 0:00:24.101 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:25:01 +0000 (0:00:00.022) 0:00:24.123 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:38 Thursday 21 July 2022 11:25:02 +0000 (0:00:01.000) 0:00:25.124 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 11:25:02 +0000 (0:00:00.041) 0:00:25.165 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 11:25:02 +0000 (0:00:00.051) 0:00:25.217 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 11:25:02 +0000 (0:00:00.036) 0:00:25.253 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "orpLHs-QAlO-ofxm-PXaX-7fbW-EoVI-SkA47O" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "GVpqOl-Yf8l-mS03-BoVV-a8c9-d3Et-5XzWy7" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "jaXvrO-mdEi-rCM1-gqIh-G5wr-UjQv-5zhYG9" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-11-24-24-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 11:25:02 +0000 (0:00:00.496) 0:00:25.750 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002321", "end": "2022-07-21 07:25:02.953344", "rc": 0, "start": "2022-07-21 07:25:02.951023" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/vg1-lv1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 11:25:03 +0000 (0:00:00.498) 0:00:26.249 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002712", "end": "2022-07-21 07:25:03.329674", "failed_when_result": false, "rc": 1, "start": "2022-07-21 07:25:03.326962" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 11:25:03 +0000 (0:00:00.372) 0:00:26.621 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 11:25:03 +0000 (0:00:00.058) 0:00:26.679 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 11:25:03 +0000 (0:00:00.031) 0:00:26.711 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 11:25:03 +0000 (0:00:00.051) 0:00:26.762 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 11:25:03 +0000 (0:00:00.088) 0:00:26.850 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 11:25:05 +0000 (0:00:01.120) 0:00:27.971 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.051) 0:00:28.022 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.060) 0:00:28.083 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.055) 0:00:28.138 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.039) 0:00:28.178 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.051) 0:00:28.230 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.028) 0:00:28.258 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.072) 0:00:28.331 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.047) 0:00:28.378 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.026) 0:00:28.405 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.026) 0:00:28.431 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.026) 0:00:28.458 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.028) 0:00:28.486 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.027) 0:00:28.514 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.027) 0:00:28.541 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.027) 0:00:28.569 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.040) 0:00:28.610 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.050) 0:00:28.660 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.051) 0:00:28.711 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.031) 0:00:28.743 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.032) 0:00:28.775 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.033) 0:00:28.808 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 11:25:05 +0000 (0:00:00.075) 0:00:28.884 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.048) 0:00:28.933 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv1&&segtype=thin", "vg1" ], "delta": "0:00:00.042433", "end": "2022-07-21 07:25:06.055855", "rc": 0, "start": "2022-07-21 07:25:06.013422" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.420) 0:00:29.353 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.057) 0:00:29.411 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.060) 0:00:29.471 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.043) 0:00:29.515 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.049) 0:00:29.565 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.051) 0:00:29.617 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.036) 0:00:29.653 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.054) 0:00:29.708 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.047) 0:00:29.756 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.052) 0:00:29.809 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.039) 0:00:29.848 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.037) 0:00:29.886 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:06 +0000 (0:00:00.042) 0:00:29.928 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.034) 0:00:29.962 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.048) 0:00:30.011 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.052) 0:00:30.063 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.038) 0:00:30.101 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.037) 0:00:30.139 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.040) 0:00:30.180 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.036) 0:00:30.216 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.053) 0:00:30.269 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.055) 0:00:30.325 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.041) 0:00:30.367 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.036) 0:00:30.403 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.037) 0:00:30.440 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.102) 0:00:30.543 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.036) 0:00:30.579 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.053) 0:00:30.633 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.049) 0:00:30.682 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.027) 0:00:30.709 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.026) 0:00:30.736 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.026) 0:00:30.763 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.026) 0:00:30.789 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.026) 0:00:30.816 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.026) 0:00:30.842 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.026) 0:00:30.869 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 11:25:07 +0000 (0:00:00.038) 0:00:30.908 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.039) 0:00:30.947 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.046) 0:00:30.993 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.053) 0:00:31.046 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.083) 0:00:31.130 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.042) 0:00:31.173 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.058) 0:00:31.232 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.053) 0:00:31.285 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.052) 0:00:31.337 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.055) 0:00:31.393 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.026) 0:00:31.419 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.026) 0:00:31.445 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.025) 0:00:31.471 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.039) 0:00:31.510 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/vg1-lv1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.062) 0:00:31.572 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.052) 0:00:31.624 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.053) 0:00:31.678 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.037) 0:00:31.716 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.069) 0:00:31.786 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 11:25:08 +0000 (0:00:00.121) 0:00:31.908 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.045) 0:00:31.953 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658402697.8663177, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658402697.8663177, "dev": 5, "device_type": 64772, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1050, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658402697.8663177, "nlink": 1, "path": "/dev/mapper/vg1-lv1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.375) 0:00:32.329 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.041) 0:00:32.370 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.039) 0:00:32.410 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.038) 0:00:32.448 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.024) 0:00:32.473 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.040) 0:00:32.513 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 11:25:09 +0000 (0:00:00.026) 0:00:32.539 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.882) 0:00:33.422 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.026) 0:00:33.448 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.024) 0:00:33.472 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.053) 0:00:33.526 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.027) 0:00:33.553 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.029) 0:00:33.583 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.026) 0:00:33.610 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.025) 0:00:33.635 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.027) 0:00:33.663 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.053) 0:00:33.717 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.052) 0:00:33.769 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.041) 0:00:33.810 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.038) 0:00:33.849 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 11:25:10 +0000 (0:00:00.039) 0:00:33.888 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.118) 0:00:34.006 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.039) 0:00:34.045 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.038) 0:00:34.084 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.039) 0:00:34.123 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.036) 0:00:34.159 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.040) 0:00:34.200 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.042) 0:00:34.242 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.039) 0:00:34.282 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 11:25:11 +0000 (0:00:00.463) 0:00:34.745 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.379) 0:00:35.125 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.054) 0:00:35.179 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.040) 0:00:35.220 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.038) 0:00:35.258 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.040) 0:00:35.299 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.039) 0:00:35.339 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.036) 0:00:35.375 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.037) 0:00:35.413 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.040) 0:00:35.453 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.036) 0:00:35.490 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 11:25:12 +0000 (0:00:00.055) 0:00:35.546 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv1" ], "delta": "0:00:00.038934", "end": "2022-07-21 07:25:12.661766", "rc": 0, "start": "2022-07-21 07:25:12.622832" } STDOUT: LVM2_LV_NAME=lv1 LVM2_LV_ATTR=Vwi-aotz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.409) 0:00:35.955 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.091) 0:00:36.047 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.131) 0:00:36.179 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.039) 0:00:36.218 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.037) 0:00:36.256 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.039) 0:00:36.296 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.038) 0:00:36.334 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.035) 0:00:36.369 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.022) 0:00:36.392 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Repeat previous invocation to verify idempotence] ************************ task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:40 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.040) 0:00:36.432 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.053) 0:00:36.486 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:25:13 +0000 (0:00:00.040) 0:00:36.526 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:25:14 +0000 (0:00:00.524) 0:00:37.050 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:25:14 +0000 (0:00:00.062) 0:00:37.113 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:25:14 +0000 (0:00:00.034) 0:00:37.147 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:25:14 +0000 (0:00:00.032) 0:00:37.179 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:25:14 +0000 (0:00:00.054) 0:00:37.234 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:25:14 +0000 (0:00:00.020) 0:00:37.254 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:25:15 +0000 (0:00:00.838) 0:00:38.093 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "3g", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:25:15 +0000 (0:00:00.041) 0:00:38.134 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:25:15 +0000 (0:00:00.076) 0:00:38.211 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:25:17 +0000 (0:00:02.542) 0:00:40.753 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:25:17 +0000 (0:00:00.056) 0:00:40.809 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:25:17 +0000 (0:00:00.039) 0:00:40.849 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:25:17 +0000 (0:00:00.042) 0:00:40.892 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:25:17 +0000 (0:00:00.037) 0:00:40.930 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:25:18 +0000 (0:00:00.838) 0:00:41.768 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:25:20 +0000 (0:00:01.665) 0:00:43.433 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:25:20 +0000 (0:00:00.058) 0:00:43.492 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:25:20 +0000 (0:00:00.022) 0:00:43.514 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "lvm2", "dosfstools", "e2fsprogs", "xfsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:25:23 +0000 (0:00:02.786) 0:00:46.300 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:25:23 +0000 (0:00:00.039) 0:00:46.340 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:25:23 +0000 (0:00:00.024) 0:00:46.365 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" } ], "packages": [ "lvm2", "dosfstools", "e2fsprogs", "xfsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:25:23 +0000 (0:00:00.040) 0:00:46.406 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:25:23 +0000 (0:00:00.076) 0:00:46.482 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:25:23 +0000 (0:00:00.074) 0:00:46.556 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:25:23 +0000 (0:00:00.040) 0:00:46.597 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:25:24 +0000 (0:00:00.719) 0:00:47.317 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount ok: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv1', 'path': '/opt/test1', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": false, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test1", "src": "/dev/mapper/vg1-lv1", "state": "mounted" }, "name": "/opt/test1", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv1" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:25:24 +0000 (0:00:00.395) 0:00:47.712 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:25:25 +0000 (0:00:00.685) 0:00:48.397 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:25:25 +0000 (0:00:00.379) 0:00:48.777 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:25:25 +0000 (0:00:00.026) 0:00:48.803 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:56 Thursday 21 July 2022 11:25:26 +0000 (0:00:00.997) 0:00:49.801 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 11:25:26 +0000 (0:00:00.046) 0:00:49.848 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 11:25:26 +0000 (0:00:00.053) 0:00:49.901 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 11:25:27 +0000 (0:00:00.038) 0:00:49.940 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "orpLHs-QAlO-ofxm-PXaX-7fbW-EoVI-SkA47O" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "GVpqOl-Yf8l-mS03-BoVV-a8c9-d3Et-5XzWy7" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "jaXvrO-mdEi-rCM1-gqIh-G5wr-UjQv-5zhYG9" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-11-24-24-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 11:25:27 +0000 (0:00:00.386) 0:00:50.326 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002408", "end": "2022-07-21 07:25:27.396303", "rc": 0, "start": "2022-07-21 07:25:27.393895" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/vg1-lv1 /opt/test1 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 11:25:27 +0000 (0:00:00.363) 0:00:50.690 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002473", "end": "2022-07-21 07:25:27.801602", "failed_when_result": false, "rc": 1, "start": "2022-07-21 07:25:27.799129" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 11:25:28 +0000 (0:00:00.406) 0:00:51.096 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 11:25:28 +0000 (0:00:00.062) 0:00:51.158 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 11:25:28 +0000 (0:00:00.103) 0:00:51.261 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 11:25:28 +0000 (0:00:00.049) 0:00:51.311 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 11:25:28 +0000 (0:00:00.063) 0:00:51.374 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 11:25:29 +0000 (0:00:01.064) 0:00:52.439 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.055) 0:00:52.494 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.051) 0:00:52.545 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.056) 0:00:52.602 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.044) 0:00:52.646 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.049) 0:00:52.695 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.025) 0:00:52.721 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.075) 0:00:52.796 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.045) 0:00:52.841 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.026) 0:00:52.868 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.026) 0:00:52.894 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 11:25:29 +0000 (0:00:00.024) 0:00:52.918 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.026) 0:00:52.945 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.032) 0:00:52.978 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.031) 0:00:53.010 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.028) 0:00:53.038 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.038) 0:00:53.077 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.047) 0:00:53.124 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.046) 0:00:53.170 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.067) 0:00:53.237 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.032) 0:00:53.270 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.031) 0:00:53.301 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.050) 0:00:53.351 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.047) 0:00:53.399 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv1&&segtype=thin", "vg1" ], "delta": "0:00:00.038799", "end": "2022-07-21 07:25:30.536578", "rc": 0, "start": "2022-07-21 07:25:30.497779" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.433) 0:00:53.832 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 11:25:30 +0000 (0:00:00.060) 0:00:53.893 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.053) 0:00:53.946 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.040) 0:00:53.987 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.050) 0:00:54.037 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.054) 0:00:54.091 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.034) 0:00:54.126 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.055) 0:00:54.181 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.050) 0:00:54.232 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.053) 0:00:54.285 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.040) 0:00:54.326 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.038) 0:00:54.365 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.040) 0:00:54.405 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.037) 0:00:54.443 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.052) 0:00:54.495 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.058) 0:00:54.554 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.038) 0:00:54.592 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.036) 0:00:54.628 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.036) 0:00:54.664 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.036) 0:00:54.700 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.048) 0:00:54.749 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:31 +0000 (0:00:00.098) 0:00:54.847 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.087) 0:00:54.934 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.040) 0:00:54.974 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.039) 0:00:55.014 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.039) 0:00:55.053 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.036) 0:00:55.090 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.051) 0:00:55.141 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.049) 0:00:55.191 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.026) 0:00:55.218 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.026) 0:00:55.245 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.025) 0:00:55.271 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.026) 0:00:55.297 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.026) 0:00:55.323 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.027) 0:00:55.351 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.025) 0:00:55.377 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.037) 0:00:55.415 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.038) 0:00:55.453 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.046) 0:00:55.500 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.050) 0:00:55.550 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.078) 0:00:55.629 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.041) 0:00:55.670 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 770083, "block_size": 4096, "block_total": 783872, "block_used": 13789, "device": "/dev/mapper/vg1-lv1", "fstype": "xfs", "inode_available": 1572861, "inode_total": 1572864, "inode_used": 3, "mount": "/opt/test1", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 3154259968, "size_total": 3210739712, "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.053) 0:00:55.724 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.050) 0:00:55.775 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.050) 0:00:55.825 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.049) 0:00:55.875 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.025) 0:00:55.901 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 11:25:32 +0000 (0:00:00.025) 0:00:55.926 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.026) 0:00:55.952 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.034) 0:00:55.986 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/vg1-lv1 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test1 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test1 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.063) 0:00:56.050 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.051) 0:00:56.101 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.052) 0:00:56.154 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.039) 0:00:56.194 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.112) 0:00:56.306 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.039) 0:00:56.346 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.041) 0:00:56.387 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658402697.8663177, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658402697.8663177, "dev": 5, "device_type": 64772, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1050, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658402697.8663177, "nlink": 1, "path": "/dev/mapper/vg1-lv1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.375) 0:00:56.763 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.043) 0:00:56.806 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.042) 0:00:56.849 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.038) 0:00:56.888 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 11:25:33 +0000 (0:00:00.024) 0:00:56.912 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 11:25:34 +0000 (0:00:00.042) 0:00:56.955 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 11:25:34 +0000 (0:00:00.025) 0:00:56.981 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.959) 0:00:57.940 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.027) 0:00:57.967 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.026) 0:00:57.994 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.055) 0:00:58.049 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.026) 0:00:58.076 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.027) 0:00:58.103 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.027) 0:00:58.130 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.027) 0:00:58.158 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.026) 0:00:58.184 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.053) 0:00:58.237 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.055) 0:00:58.293 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.038) 0:00:58.331 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.039) 0:00:58.371 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.041) 0:00:58.412 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.036) 0:00:58.448 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.040) 0:00:58.489 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.041) 0:00:58.530 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.079) 0:00:58.609 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.039) 0:00:58.649 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.038) 0:00:58.688 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.042) 0:00:58.730 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 11:25:35 +0000 (0:00:00.039) 0:00:58.770 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.367) 0:00:59.137 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.375) 0:00:59.513 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.050) 0:00:59.563 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.036) 0:00:59.599 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.034) 0:00:59.634 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.034) 0:00:59.669 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.035) 0:00:59.705 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.034) 0:00:59.740 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.034) 0:00:59.774 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.036) 0:00:59.811 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.033) 0:00:59.845 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 11:25:36 +0000 (0:00:00.048) 0:00:59.894 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv1" ], "delta": "0:00:00.033771", "end": "2022-07-21 07:25:36.998324", "rc": 0, "start": "2022-07-21 07:25:36.964553" } STDOUT: LVM2_LV_NAME=lv1 LVM2_LV_ATTR=Vwi-aotz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.400) 0:01:00.294 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.049) 0:01:00.343 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.052) 0:01:00.395 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.038) 0:01:00.434 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.037) 0:01:00.472 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.037) 0:01:00.509 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.038) 0:01:00.548 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.072) 0:01:00.620 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.024) 0:01:00.645 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Change thinlv fs type] *************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:58 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.035) 0:01:00.681 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.056) 0:01:00.738 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:25:37 +0000 (0:00:00.038) 0:01:00.777 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:25:38 +0000 (0:00:00.530) 0:01:01.307 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:25:38 +0000 (0:00:00.063) 0:01:01.371 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:25:38 +0000 (0:00:00.040) 0:01:01.411 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:25:38 +0000 (0:00:00.043) 0:01:01.454 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:25:38 +0000 (0:00:00.058) 0:01:01.513 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:25:38 +0000 (0:00:00.024) 0:01:01.537 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:25:39 +0000 (0:00:00.866) 0:01:02.403 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "fs_type": "xfs", "name": "lv1", "thin": true, "thin_pool_name": "tpool1" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:25:39 +0000 (0:00:00.054) 0:01:02.458 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:25:39 +0000 (0:00:00.038) 0:01:02.496 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2", "xfsprogs" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:25:42 +0000 (0:00:02.659) 0:01:05.156 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:25:42 +0000 (0:00:00.051) 0:01:05.207 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:25:42 +0000 (0:00:00.038) 0:01:05.245 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:25:42 +0000 (0:00:00.041) 0:01:05.287 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:25:42 +0000 (0:00:00.040) 0:01:05.328 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:25:43 +0000 (0:00:00.856) 0:01:06.185 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:25:45 +0000 (0:00:01.745) 0:01:07.931 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:25:45 +0000 (0:00:00.090) 0:01:08.022 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:25:45 +0000 (0:00:00.024) 0:01:08.046 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "path": "/opt/test1", "state": "absent" } ], "packages": [ "lvm2", "xfsprogs", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:25:47 +0000 (0:00:02.688) 0:01:10.734 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:25:47 +0000 (0:00:00.039) 0:01:10.774 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:25:47 +0000 (0:00:00.022) 0:01:10.797 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [], "changed": false, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "path": "/opt/test1", "state": "absent" } ], "packages": [ "lvm2", "xfsprogs", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:25:47 +0000 (0:00:00.073) 0:01:10.870 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:25:48 +0000 (0:00:00.073) 0:01:10.944 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:25:48 +0000 (0:00:00.036) 0:01:10.981 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'path': '/opt/test1', 'state': 'absent'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "mount_info": { "path": "/opt/test1", "state": "absent" }, "name": "/opt/test1", "opts": "defaults", "passno": "0" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:25:48 +0000 (0:00:00.416) 0:01:11.397 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:25:49 +0000 (0:00:00.675) 0:01:12.072 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:25:49 +0000 (0:00:00.043) 0:01:12.115 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:25:49 +0000 (0:00:00.671) 0:01:12.787 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:25:50 +0000 (0:00:00.369) 0:01:13.156 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:25:50 +0000 (0:00:00.023) 0:01:13.179 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:72 Thursday 21 July 2022 11:25:51 +0000 (0:00:01.042) 0:01:14.222 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 11:25:51 +0000 (0:00:00.050) 0:01:14.272 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_kernel_device": "/dev/dm-4", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "_raw_kernel_device": "/dev/dm-4", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": null, "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 3221225472, "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 11:25:51 +0000 (0:00:00.055) 0:01:14.328 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 11:25:51 +0000 (0:00:00.044) 0:01:14.373 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "orpLHs-QAlO-ofxm-PXaX-7fbW-EoVI-SkA47O" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "GVpqOl-Yf8l-mS03-BoVV-a8c9-d3Et-5XzWy7" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "jaXvrO-mdEi-rCM1-gqIh-G5wr-UjQv-5zhYG9" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-11-24-24-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 11:25:51 +0000 (0:00:00.383) 0:01:14.757 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002721", "end": "2022-07-21 07:25:51.891368", "rc": 0, "start": "2022-07-21 07:25:51.888647" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 11:25:52 +0000 (0:00:00.430) 0:01:15.188 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002435", "end": "2022-07-21 07:25:52.302347", "failed_when_result": false, "rc": 1, "start": "2022-07-21 07:25:52.299912" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 11:25:52 +0000 (0:00:00.409) 0:01:15.597 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 11:25:52 +0000 (0:00:00.061) 0:01:15.659 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 11:25:52 +0000 (0:00:00.034) 0:01:15.693 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 11:25:52 +0000 (0:00:00.046) 0:01:15.740 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 11:25:52 +0000 (0:00:00.060) 0:01:15.800 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 11:25:53 +0000 (0:00:01.072) 0:01:16.873 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 11:25:53 +0000 (0:00:00.053) 0:01:16.926 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.051) 0:01:16.978 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.053) 0:01:17.031 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.038) 0:01:17.069 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.053) 0:01:17.123 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.026) 0:01:17.149 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.073) 0:01:17.223 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.044) 0:01:17.267 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.028) 0:01:17.295 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.026) 0:01:17.322 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.027) 0:01:17.349 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.027) 0:01:17.377 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.024) 0:01:17.401 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.023) 0:01:17.425 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.023) 0:01:17.449 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.098) 0:01:17.547 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.052) 0:01:17.600 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.046) 0:01:17.646 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.029) 0:01:17.676 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.029) 0:01:17.706 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.030) 0:01:17.736 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.045) 0:01:17.782 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 11:25:54 +0000 (0:00:00.046) 0:01:17.828 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv1&&segtype=thin", "vg1" ], "delta": "0:00:00.037511", "end": "2022-07-21 07:25:54.953336", "rc": 0, "start": "2022-07-21 07:25:54.915825" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.421) 0:01:18.250 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.057) 0:01:18.307 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.054) 0:01:18.362 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.045) 0:01:18.407 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.049) 0:01:18.457 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.054) 0:01:18.511 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.032) 0:01:18.544 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.056) 0:01:18.600 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.052) 0:01:18.652 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.051) 0:01:18.703 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.052) 0:01:18.755 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.039) 0:01:18.795 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.038) 0:01:18.833 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.040) 0:01:18.873 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:55 +0000 (0:00:00.053) 0:01:18.927 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.053) 0:01:18.981 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.038) 0:01:19.019 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.036) 0:01:19.056 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.036) 0:01:19.092 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.087) 0:01:19.179 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.047) 0:01:19.227 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.048) 0:01:19.275 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.039) 0:01:19.315 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.037) 0:01:19.352 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.040) 0:01:19.393 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.035) 0:01:19.429 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.038) 0:01:19.467 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.049) 0:01:19.517 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.048) 0:01:19.565 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.025) 0:01:19.591 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.024) 0:01:19.616 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.025) 0:01:19.642 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.025) 0:01:19.667 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.026) 0:01:19.693 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.024) 0:01:19.717 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.025) 0:01:19.743 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.038) 0:01:19.782 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.040) 0:01:19.822 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': None, 'name': 'lv1', 'raid_level': None, 'size': 3221225472, 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1', '_kernel_device': '/dev/dm-4', '_raw_kernel_device': '/dev/dm-4'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.044) 0:01:19.867 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 11:25:56 +0000 (0:00:00.056) 0:01:19.924 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.079) 0:01:20.003 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.045) 0:01:20.049 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.061) 0:01:20.110 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.027) 0:01:20.137 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.051) 0:01:20.189 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.037) 0:01:20.226 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.023) 0:01:20.250 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.024) 0:01:20.274 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.025) 0:01:20.300 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.078) 0:01:20.378 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.155) 0:01:20.534 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.051) 0:01:20.586 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.050) 0:01:20.636 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.041) 0:01:20.678 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.037) 0:01:20.716 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.040) 0:01:20.756 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 11:25:57 +0000 (0:00:00.042) 0:01:20.799 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658402697.8663177, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658402697.8663177, "dev": 5, "device_type": 64772, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1050, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658402697.8663177, "nlink": 1, "path": "/dev/mapper/vg1-lv1", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.369) 0:01:21.168 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.042) 0:01:21.210 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.038) 0:01:21.249 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.036) 0:01:21.286 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.024) 0:01:21.310 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.040) 0:01:21.350 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 11:25:58 +0000 (0:00:00.023) 0:01:21.374 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.900) 0:01:22.275 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.025) 0:01:22.300 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.027) 0:01:22.327 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.054) 0:01:22.381 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.025) 0:01:22.407 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.025) 0:01:22.432 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.025) 0:01:22.458 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.024) 0:01:22.482 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.024) 0:01:22.507 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.054) 0:01:22.561 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.051) 0:01:22.613 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.038) 0:01:22.651 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.039) 0:01:22.691 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.072) 0:01:22.763 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.034) 0:01:22.797 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.037) 0:01:22.835 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.041) 0:01:22.876 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 11:25:59 +0000 (0:00:00.037) 0:01:22.914 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.038) 0:01:22.953 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.042) 0:01:22.995 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.037) 0:01:23.032 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.039) 0:01:23.072 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.383) 0:01:23.456 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 3221225472, "changed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.385) 0:01:23.841 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "3221225472" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 11:26:00 +0000 (0:00:00.053) 0:01:23.895 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.035) 0:01:23.930 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.038) 0:01:23.969 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.038) 0:01:24.008 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.038) 0:01:24.046 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.036) 0:01:24.083 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.037) 0:01:24.121 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 3221225472, "changed": false, "failed": false, "lvm": "3g", "parted": "3GiB", "size": "3 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.037) 0:01:24.158 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "3221225472" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.038) 0:01:24.196 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.058) 0:01:24.255 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv1" ], "delta": "0:00:00.037847", "end": "2022-07-21 07:26:01.391330", "rc": 0, "start": "2022-07-21 07:26:01.353483" } STDOUT: LVM2_LV_NAME=lv1 LVM2_LV_ATTR=Vwi-a-tz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.438) 0:01:24.693 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.082) 0:01:24.775 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.094) 0:01:24.869 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 11:26:01 +0000 (0:00:00.042) 0:01:24.912 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.042) 0:01:24.954 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.039) 0:01:24.994 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.041) 0:01:25.035 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.039) 0:01:25.075 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.022) 0:01:25.097 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Create new LV under existing thinpool] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:74 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.035) 0:01:25.133 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.062) 0:01:25.195 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.037) 0:01:25.233 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.523) 0:01:25.756 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.068) 0:01:25.825 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.042) 0:01:25.867 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:26:02 +0000 (0:00:00.043) 0:01:25.910 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:26:03 +0000 (0:00:00.058) 0:01:25.969 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:26:03 +0000 (0:00:00.022) 0:01:25.991 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:26:03 +0000 (0:00:00.889) 0:01:26.881 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "mount_point": "/opt/test2", "name": "lv2", "size": "4g", "thin": true, "thin_pool_name": "tpool1" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:26:04 +0000 (0:00:00.078) 0:01:26.960 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:26:04 +0000 (0:00:00.036) 0:01:26.997 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:26:06 +0000 (0:00:02.753) 0:01:29.750 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:26:06 +0000 (0:00:00.053) 0:01:29.804 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:26:06 +0000 (0:00:00.038) 0:01:29.842 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:26:06 +0000 (0:00:00.041) 0:01:29.884 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:26:06 +0000 (0:00:00.040) 0:01:29.925 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:26:07 +0000 (0:00:00.866) 0:01:30.791 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:26:09 +0000 (0:00:01.796) 0:01:32.588 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:26:09 +0000 (0:00:00.058) 0:01:32.646 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:26:09 +0000 (0:00:00.022) 0:01:32.669 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "create device", "device": "/dev/mapper/vg1-lv2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "mounted" } ], "packages": [ "xfsprogs", "e2fsprogs", "dosfstools", "lvm2" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:26:12 +0000 (0:00:03.038) 0:01:35.707 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:26:12 +0000 (0:00:00.036) 0:01:35.744 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:26:12 +0000 (0:00:00.022) 0:01:35.767 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "create device", "device": "/dev/mapper/vg1-lv2", "fs_type": null }, { "action": "create format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf", "/dev/mapper/vg1-lv2" ], "mounts": [ { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "mounted" } ], "packages": [ "xfsprogs", "e2fsprogs", "dosfstools", "lvm2" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:26:12 +0000 (0:00:00.074) 0:01:35.841 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:26:12 +0000 (0:00:00.043) 0:01:35.884 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:26:12 +0000 (0:00:00.042) 0:01:35.927 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:26:13 +0000 (0:00:00.041) 0:01:35.968 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:26:13 +0000 (0:00:00.698) 0:01:36.667 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv2', 'path': '/opt/test2', 'fstype': 'xfs', 'opts': 'defaults', 'dump': 0, 'passno': 0, 'state': 'mounted'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "dump": 0, "fstype": "xfs", "opts": "defaults", "passno": 0, "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "mounted" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:26:14 +0000 (0:00:00.433) 0:01:37.100 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:26:14 +0000 (0:00:00.687) 0:01:37.788 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:26:15 +0000 (0:00:00.377) 0:01:38.165 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:26:15 +0000 (0:00:00.025) 0:01:38.191 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:89 Thursday 21 July 2022 11:26:16 +0000 (0:00:01.015) 0:01:39.207 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 11:26:16 +0000 (0:00:00.053) 0:01:39.260 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_kernel_device": "/dev/dm-5", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "_raw_kernel_device": "/dev/dm-5", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "4g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 11:26:16 +0000 (0:00:00.083) 0:01:39.343 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 11:26:16 +0000 (0:00:00.037) 0:01:39.380 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" }, "/dev/mapper/vg1-lv2": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv2", "size": "4G", "type": "lvm", "uuid": "bc2861f7-b9cc-46fe-8176-fdbd67ea1b88" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "orpLHs-QAlO-ofxm-PXaX-7fbW-EoVI-SkA47O" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "GVpqOl-Yf8l-mS03-BoVV-a8c9-d3Et-5XzWy7" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "jaXvrO-mdEi-rCM1-gqIh-G5wr-UjQv-5zhYG9" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-11-24-24-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 11:26:16 +0000 (0:00:00.404) 0:01:39.785 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002495", "end": "2022-07-21 07:26:16.860847", "rc": 0, "start": "2022-07-21 07:26:16.858352" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/mapper/vg1-lv2 /opt/test2 xfs defaults 0 0 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 11:26:17 +0000 (0:00:00.371) 0:01:40.156 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.004275", "end": "2022-07-21 07:26:18.232948", "failed_when_result": false, "rc": 1, "start": "2022-07-21 07:26:17.228673" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 11:26:18 +0000 (0:00:01.382) 0:01:41.538 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 11:26:18 +0000 (0:00:00.063) 0:01:41.602 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 11:26:18 +0000 (0:00:00.035) 0:01:41.637 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 11:26:18 +0000 (0:00:00.047) 0:01:41.684 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 11:26:18 +0000 (0:00:00.056) 0:01:41.741 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 11:26:19 +0000 (0:00:01.136) 0:01:42.878 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.054) 0:01:42.933 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.084) 0:01:43.017 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.080) 0:01:43.098 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.038) 0:01:43.136 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.081) 0:01:43.218 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.052) 0:01:43.270 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.070) 0:01:43.341 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.043) 0:01:43.384 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.023) 0:01:43.408 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.024) 0:01:43.432 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.025) 0:01:43.457 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.024) 0:01:43.481 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.023) 0:01:43.505 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.028) 0:01:43.534 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.027) 0:01:43.561 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.038) 0:01:43.599 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.049) 0:01:43.649 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.046) 0:01:43.695 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.034) 0:01:43.729 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.030) 0:01:43.760 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.030) 0:01:43.790 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.048) 0:01:43.839 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 11:26:20 +0000 (0:00:00.046) 0:01:43.885 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheading", "-o", "pool_lv", "--select", "lv_name=lv2&&segtype=thin", "vg1" ], "delta": "0:00:00.039820", "end": "2022-07-21 07:26:21.022519", "rc": 0, "start": "2022-07-21 07:26:20.982699" } STDOUT: tpool1 TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.437) 0:01:44.323 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.060) 0:01:44.383 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.054) 0:01:44.437 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lvmraid_status": null }, "changed": false } TASK [Check member encryption] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.044) 0:01:44.482 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.049) 0:01:44.531 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.126) 0:01:44.658 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.033) 0:01:44.692 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.054) 0:01:44.746 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.050) 0:01:44.796 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.054) 0:01:44.850 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.038) 0:01:44.889 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:26:21 +0000 (0:00:00.039) 0:01:44.928 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.037) 0:01:44.966 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.056) 0:01:45.022 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.052) 0:01:45.074 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.054) 0:01:45.129 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.038) 0:01:45.168 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.037) 0:01:45.205 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.041) 0:01:45.247 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.034) 0:01:45.281 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.054) 0:01:45.336 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.051) 0:01:45.388 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.038) 0:01:45.426 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.040) 0:01:45.466 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.039) 0:01:45.506 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.038) 0:01:45.544 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.041) 0:01:45.586 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.051) 0:01:45.637 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.047) 0:01:45.684 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.024) 0:01:45.709 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.024) 0:01:45.734 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.023) 0:01:45.757 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.024) 0:01:45.781 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.026) 0:01:45.808 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.064) 0:01:45.873 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 11:26:22 +0000 (0:00:00.027) 0:01:45.901 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.038) 0:01:45.939 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.037) 0:01:45.977 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': '4g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': [], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2', '_kernel_device': '/dev/dm-5', '_raw_kernel_device': '/dev/dm-5'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.044) 0:01:46.021 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": true, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.049) 0:01:46.071 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.076) 0:01:46.147 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.045) 0:01:46.193 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [ { "block_available": 1030395, "block_size": 4096, "block_total": 1046016, "block_used": 15621, "device": "/dev/mapper/vg1-lv2", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 4220497920, "size_total": 4284481536, "uuid": "bc2861f7-b9cc-46fe-8176-fdbd67ea1b88" } ], "storage_test_mount_expected_match_count": "1", "storage_test_mount_point_matches": [ { "block_available": 1030395, "block_size": 4096, "block_total": 1046016, "block_used": 15621, "device": "/dev/mapper/vg1-lv2", "fstype": "xfs", "inode_available": 2097149, "inode_total": 2097152, "inode_used": 3, "mount": "/opt/test2", "options": "rw,seclabel,relatime,attr2,inode64,logbufs=8,logbsize=64k,sunit=128,swidth=128,noquota", "size_available": 4220497920, "size_total": 4284481536, "uuid": "bc2861f7-b9cc-46fe-8176-fdbd67ea1b88" } ], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.062) 0:01:46.255 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.053) 0:01:46.309 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.069) 0:01:46.379 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [command] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.055) 0:01:46.434 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.025) 0:01:46.459 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.026) 0:01:46.486 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.025) 0:01:46.511 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.038) 0:01:46.549 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "1", "storage_test_fstab_expected_mount_options_matches": "1", "storage_test_fstab_expected_mount_point_matches": "1", "storage_test_fstab_id_matches": [ "/dev/mapper/vg1-lv2 " ], "storage_test_fstab_mount_options_matches": [ " /opt/test2 xfs defaults " ], "storage_test_fstab_mount_point_matches": [ " /opt/test2 " ] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.066) 0:01:46.616 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.052) 0:01:46.668 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.054) 0:01:46.723 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.040) 0:01:46.764 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.035) 0:01:46.799 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify fs label] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.038) 0:01:46.838 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [See whether the device node is present] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 11:26:23 +0000 (0:00:00.041) 0:01:46.879 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "atime": 1658402772.3433177, "attr_flags": "", "attributes": [], "block_size": 4096, "blocks": 0, "charset": "binary", "ctime": 1658402772.3433177, "dev": 5, "device_type": 64773, "executable": false, "exists": true, "gid": 6, "gr_name": "disk", "inode": 1632, "isblk": true, "ischr": false, "isdir": false, "isfifo": false, "isgid": false, "islnk": false, "isreg": false, "issock": false, "isuid": false, "mimetype": "inode/symlink", "mode": "0660", "mtime": 1658402772.3433177, "nlink": 1, "path": "/dev/mapper/vg1-lv2", "pw_name": "root", "readable": true, "rgrp": true, "roth": false, "rusr": true, "size": 0, "uid": 0, "version": null, "wgrp": true, "woth": false, "writeable": true, "wusr": true, "xgrp": false, "xoth": false, "xusr": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.375) 0:01:47.255 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.085) 0:01:47.341 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.085) 0:01:47.426 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.036) 0:01:47.463 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.025) 0:01:47.488 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.037) 0:01:47.525 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 11:26:24 +0000 (0:00:00.023) 0:01:47.548 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.900) 0:01:48.449 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.025) 0:01:48.475 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.024) 0:01:48.499 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.056) 0:01:48.556 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.025) 0:01:48.581 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.025) 0:01:48.606 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.027) 0:01:48.634 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.029) 0:01:48.663 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.025) 0:01:48.689 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.052) 0:01:48.742 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.055) 0:01:48.797 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.041) 0:01:48.838 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.039) 0:01:48.878 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 11:26:25 +0000 (0:00:00.040) 0:01:48.919 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.037) 0:01:48.956 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.039) 0:01:48.996 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.041) 0:01:49.038 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.039) 0:01:49.078 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.038) 0:01:49.116 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.040) 0:01:49.156 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.074) 0:01:49.230 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.040) 0:01:49.271 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 11:26:26 +0000 (0:00:00.374) 0:01:49.645 ********* ok: [/cache/rhel-9.qcow2.snap] => { "bytes": 4294967296, "changed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.381) 0:01:50.026 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_expected_size": "4294967296" }, "changed": false } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.054) 0:01:50.081 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.037) 0:01:50.118 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.038) 0:01:50.157 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.042) 0:01:50.199 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.041) 0:01:50.240 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.039) 0:01:50.279 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.042) 0:01:50.321 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "bytes": 4294967296, "changed": false, "failed": false, "lvm": "4g", "parted": "4GiB", "size": "4 GiB" } } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.039) 0:01:50.361 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.040) 0:01:50.401 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Get information about the LV] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.059) 0:01:50.460 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "lvs", "--noheadings", "--nameprefixes", "--units=b", "--nosuffix", "--unquoted", "-o", "name,attr,cache_total_blocks,chunk_size,segtype", "vg1/lv2" ], "delta": "0:00:00.034223", "end": "2022-07-21 07:26:27.582783", "rc": 0, "start": "2022-07-21 07:26:27.548560" } STDOUT: LVM2_LV_NAME=lv2 LVM2_LV_ATTR=Vwi-aotz-- LVM2_CACHE_TOTAL_BLOCKS= LVM2_CHUNK_SIZE=0 LVM2_SEGTYPE=thin TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 11:26:27 +0000 (0:00:00.419) 0:01:50.880 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_lv_segtype": [ "thin" ] }, "changed": false } TASK [check segment type] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.055) 0:01:50.935 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.053) 0:01:50.988 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.042) 0:01:51.030 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.042) 0:01:51.073 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.041) 0:01:51.115 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.039) 0:01:51.155 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.112) 0:01:51.267 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.024) 0:01:51.291 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Remove existing LV under existing thinpool] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:91 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.038) 0:01:51.329 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.071) 0:01:51.401 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:26:28 +0000 (0:00:00.041) 0:01:51.442 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:26:29 +0000 (0:00:00.520) 0:01:51.962 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:26:29 +0000 (0:00:00.081) 0:01:52.044 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:26:29 +0000 (0:00:00.038) 0:01:52.082 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:26:29 +0000 (0:00:00.038) 0:01:52.121 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:26:29 +0000 (0:00:00.061) 0:01:52.182 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:26:29 +0000 (0:00:00.023) 0:01:52.205 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:26:30 +0000 (0:00:00.866) 0:01:53.072 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "type": "lvm", "volumes": [ { "mount_point": "/opt/test2", "name": "lv2", "state": "absent", "thin": true, "thin_pool_name": "tpool1" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:26:30 +0000 (0:00:00.043) 0:01:53.115 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:26:30 +0000 (0:00:00.042) 0:01:53.158 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [ "lvm2" ], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:26:33 +0000 (0:00:02.772) 0:01:55.930 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:26:33 +0000 (0:00:00.051) 0:01:55.982 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:26:33 +0000 (0:00:00.037) 0:01:56.020 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:26:33 +0000 (0:00:00.042) 0:01:56.063 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:26:33 +0000 (0:00:00.037) 0:01:56.100 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:26:34 +0000 (0:00:00.852) 0:01:56.953 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:26:35 +0000 (0:00:01.680) 0:01:58.634 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:26:35 +0000 (0:00:00.095) 0:01:58.729 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:26:35 +0000 (0:00:00.026) 0:01:58.755 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv2", "fs_type": null } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "absent" } ], "packages": [ "xfsprogs", "lvm2", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:26:39 +0000 (0:00:03.501) 0:02:02.256 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:26:39 +0000 (0:00:00.043) 0:02:02.300 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:26:39 +0000 (0:00:00.025) 0:02:02.326 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv2", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv2", "fs_type": null } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/mapper/vg1-lv1", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [ { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "absent" } ], "packages": [ "xfsprogs", "lvm2", "dosfstools", "e2fsprogs" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:26:39 +0000 (0:00:00.044) 0:02:02.370 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:26:39 +0000 (0:00:00.044) 0:02:02.415 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:26:39 +0000 (0:00:00.044) 0:02:02.459 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount changed: [/cache/rhel-9.qcow2.snap] => (item={'src': '/dev/mapper/vg1-lv2', 'path': '/opt/test2', 'state': 'absent', 'fstype': 'xfs'}) => { "ansible_loop_var": "mount_info", "backup_file": "", "boot": "yes", "changed": true, "dump": "0", "fstab": "/etc/fstab", "fstype": "xfs", "mount_info": { "fstype": "xfs", "path": "/opt/test2", "src": "/dev/mapper/vg1-lv2", "state": "absent" }, "name": "/opt/test2", "opts": "defaults", "passno": "0", "src": "/dev/mapper/vg1-lv2" } TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:26:39 +0000 (0:00:00.425) 0:02:02.884 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:26:40 +0000 (0:00:00.688) 0:02:03.573 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:26:40 +0000 (0:00:00.041) 0:02:03.614 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "name": null, "status": {} } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:26:41 +0000 (0:00:00.690) 0:02:04.304 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:26:41 +0000 (0:00:00.393) 0:02:04.698 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:26:41 +0000 (0:00:00.027) 0:02:04.725 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:106 Thursday 21 July 2022 11:26:42 +0000 (0:00:01.043) 0:02:05.769 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 11:26:42 +0000 (0:00:00.055) 0:02:05.825 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "present", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv2", "_mount_id": "/dev/mapper/vg1-lv2", "_raw_device": "/dev/mapper/vg1-lv2", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test2", "name": "lv2", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": 4294967296, "state": "absent", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": null, "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 11:26:43 +0000 (0:00:00.145) 0:02:05.970 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 11:26:43 +0000 (0:00:00.041) 0:02:06.011 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/mapper/vg1-lv1": { "fstype": "xfs", "label": "", "name": "/dev/mapper/vg1-lv1", "size": "3G", "type": "lvm", "uuid": "5138dd2f-8e3a-4372-a8de-6d269e42b347" }, "/dev/mapper/vg1-tpool1": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1-tpool": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1-tpool", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tdata": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tdata", "size": "10G", "type": "lvm", "uuid": "" }, "/dev/mapper/vg1-tpool1_tmeta": { "fstype": "", "label": "", "name": "/dev/mapper/vg1-tpool1_tmeta", "size": "12M", "type": "lvm", "uuid": "" }, "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sda1", "size": "10G", "type": "partition", "uuid": "orpLHs-QAlO-ofxm-PXaX-7fbW-EoVI-SkA47O" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdb1", "size": "10G", "type": "partition", "uuid": "GVpqOl-Yf8l-mS03-BoVV-a8c9-d3Et-5XzWy7" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc1": { "fstype": "LVM2_member", "label": "", "name": "/dev/sdc1", "size": "10G", "type": "partition", "uuid": "jaXvrO-mdEi-rCM1-gqIh-G5wr-UjQv-5zhYG9" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-11-24-24-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 11:26:43 +0000 (0:00:00.394) 0:02:06.406 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:00.002637", "end": "2022-07-21 07:26:43.479008", "rc": 0, "start": "2022-07-21 07:26:43.476371" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 11:26:43 +0000 (0:00:00.368) 0:02:06.774 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:01.003802", "end": "2022-07-21 07:26:44.848893", "failed_when_result": false, "rc": 1, "start": "2022-07-21 07:26:43.845091" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 11:26:45 +0000 (0:00:01.372) 0:02:08.147 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'present', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 11:26:45 +0000 (0:00:00.063) 0:02:08.210 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 11:26:45 +0000 (0:00:00.037) 0:02:08.248 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 11:26:45 +0000 (0:00:00.049) 0:02:08.298 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "3", "_storage_test_pool_pvs_lvm": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 11:26:45 +0000 (0:00:00.057) 0:02:08.355 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sda1", "pv": "/dev/sda1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdb1", "pv": "/dev/sdb1" } ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "device": "/dev/sdc1", "pv": "/dev/sdc1" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 11:26:46 +0000 (0:00:01.272) 0:02:09.627 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "3" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 11:26:46 +0000 (0:00:00.052) 0:02:09.680 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [ "/dev/sda1", "/dev/sdb1", "/dev/sdc1" ] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 11:26:46 +0000 (0:00:00.085) 0:02:09.765 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 11:26:46 +0000 (0:00:00.054) 0:02:09.820 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 11:26:46 +0000 (0:00:00.040) 0:02:09.860 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.087) 0:02:09.948 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.025) 0:02:09.974 ********* ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sda1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdb1" } MSG: All assertions passed ok: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "ansible_loop_var": "pv", "changed": false, "pv": "/dev/sdc1" } MSG: All assertions passed TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.099) 0:02:10.074 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.071) 0:02:10.145 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.024) 0:02:10.170 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.026) 0:02:10.196 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.026) 0:02:10.222 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.027) 0:02:10.249 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.026) 0:02:10.276 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.025) 0:02:10.302 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.025) 0:02:10.327 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.040) 0:02:10.367 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.046) 0:02:10.414 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.046) 0:02:10.461 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.037) 0:02:10.498 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.033) 0:02:10.531 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.035) 0:02:10.567 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.046) 0:02:10.613 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.046) 0:02:10.660 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.029) 0:02:10.689 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.027) 0:02:10.717 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.030) 0:02:10.747 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.027) 0:02:10.774 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.046) 0:02:10.821 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.050) 0:02:10.872 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sda1) => { "_storage_test_pool_member_path": "/dev/sda1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdb1) => { "_storage_test_pool_member_path": "/dev/sdb1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=/dev/sdc1) => { "_storage_test_pool_member_path": "/dev/sdc1", "ansible_loop_var": "_storage_test_pool_member_path", "changed": false, "skip_reason": "Conditional result was False" } TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 11:26:47 +0000 (0:00:00.030) 0:02:10.902 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sda1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdb1) included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml for /cache/rhel-9.qcow2.snap => (item=/dev/sdc1) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.050) 0:02:10.953 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.047) 0:02:11.001 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.048) 0:02:11.050 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.037) 0:02:11.087 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.036) 0:02:11.123 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.037) 0:02:11.161 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.042) 0:02:11.204 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.054) 0:02:11.259 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.107) 0:02:11.367 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.080) 0:02:11.448 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.040) 0:02:11.488 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.039) 0:02:11.527 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:1 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.050) 0:02:11.578 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [] }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:4 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.053) 0:02:11.631 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:9 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.051) 0:02:11.683 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:15 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.041) 0:02:11.724 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:21 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.039) 0:02:11.764 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-crypttab.yml:27 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.039) 0:02:11.803 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.041) 0:02:11.844 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 11:26:48 +0000 (0:00:00.035) 0:02:11.880 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.050) 0:02:11.931 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.048) 0:02:11.979 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.026) 0:02:12.005 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.025) 0:02:12.031 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.026) 0:02:12.057 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.026) 0:02:12.083 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.026) 0:02:12.110 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.027) 0:02:12.138 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.025) 0:02:12.164 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.035) 0:02:12.199 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.036) 0:02:12.236 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test2', 'name': 'lv2', 'raid_level': None, 'size': 4294967296, 'state': 'absent', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': None, 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv2', '_raw_device': '/dev/mapper/vg1-lv2', '_mount_id': '/dev/mapper/vg1-lv2'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.050) 0:02:12.286 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.057) 0:02:12.343 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.076) 0:02:12.420 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv2" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.047) 0:02:12.467 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.055) 0:02:12.522 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.023) 0:02:12.546 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.049) 0:02:12.596 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.042) 0:02:12.638 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.027) 0:02:12.665 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.026) 0:02:12.691 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.027) 0:02:12.718 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.088) 0:02:12.807 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.064) 0:02:12.872 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 11:26:49 +0000 (0:00:00.027) 0:02:12.899 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.051) 0:02:12.951 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.038) 0:02:12.990 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.040) 0:02:13.030 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.026) 0:02:13.057 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.025) 0:02:13.083 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.375) 0:02:13.458 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.042) 0:02:13.501 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.027) 0:02:13.528 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.040) 0:02:13.568 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.027) 0:02:13.596 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.025) 0:02:13.622 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 11:26:50 +0000 (0:00:00.024) 0:02:13.647 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.903) 0:02:14.550 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.025) 0:02:14.575 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.023) 0:02:14.599 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.022) 0:02:14.622 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.025) 0:02:14.647 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.028) 0:02:14.676 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.028) 0:02:14.705 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.026) 0:02:14.731 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.025) 0:02:14.757 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 11:26:51 +0000 (0:00:00.089) 0:02:14.847 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.086) 0:02:14.934 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.036) 0:02:14.971 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.067) 0:02:15.038 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.038) 0:02:15.076 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.039) 0:02:15.116 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.038) 0:02:15.155 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.037) 0:02:15.192 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.036) 0:02:15.228 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.035) 0:02:15.264 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.034) 0:02:15.299 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.034) 0:02:15.334 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.036) 0:02:15.370 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.023) 0:02:15.394 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.034) 0:02:15.429 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.034) 0:02:15.463 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.036) 0:02:15.500 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.034) 0:02:15.535 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.036) 0:02:15.571 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.040) 0:02:15.612 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.035) 0:02:15.647 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.034) 0:02:15.682 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.038) 0:02:15.720 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.038) 0:02:15.759 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.024) 0:02:15.783 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.024) 0:02:15.807 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.022) 0:02:15.830 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.022) 0:02:15.853 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.026) 0:02:15.879 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 11:26:52 +0000 (0:00:00.028) 0:02:15.908 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.026) 0:02:15.934 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.026) 0:02:15.961 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.045) 0:02:16.006 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.028) 0:02:16.035 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } TASK [Cleanup] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:108 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.038) 0:02:16.073 ********* TASK [fedora.linux_system_roles.storage : set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:2 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.078) 0:02:16.152 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : Ensure ansible_facts used by role] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:2 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.091) 0:02:16.243 ********* ok: [/cache/rhel-9.qcow2.snap] TASK [fedora.linux_system_roles.storage : Set platform/version specific variables] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/set_vars.yml:7 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.563) 0:02:16.806 ********* skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat.yml", "skip_reason": "Conditional result was False" } ok: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.yml) => { "ansible_facts": { "blivet_package_list": [ "python3-blivet", "libblockdev-crypto", "libblockdev-dm", "libblockdev-lvm", "libblockdev-mdraid", "libblockdev-swap", "vdo", "kmod-kvdo", "xfsprogs" ] }, "ansible_included_var_files": [ "/tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/vars/RedHat_9.yml" ], "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.yml" } skipping: [/cache/rhel-9.qcow2.snap] => (item=RedHat_9.0.yml) => { "ansible_loop_var": "item", "changed": false, "item": "RedHat_9.0.yml", "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : define an empty list of pools to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:5 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.068) 0:02:16.875 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : define an empty list of volumes to be used in testing] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:9 Thursday 21 July 2022 11:26:53 +0000 (0:00:00.038) 0:02:16.914 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : include the appropriate provider tasks] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main.yml:13 Thursday 21 July 2022 11:26:54 +0000 (0:00:00.039) 0:02:16.953 ********* redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount redirecting (type: modules) ansible.builtin.mount to ansible.posix.mount included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : get a list of rpm packages installed on host machine] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:2 Thursday 21 July 2022 11:26:54 +0000 (0:00:00.056) 0:02:17.010 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : make sure blivet is available] ******* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:7 Thursday 21 July 2022 11:26:54 +0000 (0:00:00.023) 0:02:17.033 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : show storage_pools] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:13 Thursday 21 July 2022 11:26:55 +0000 (0:00:00.916) 0:02:17.950 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_pools": [ { "disks": [ "sda", "sdb", "sdc" ], "name": "vg1", "state": "absent", "type": "lvm", "volumes": [ { "mount_point": "/opt/test1", "name": "lv1", "size": "3g", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g" } ] } ] } TASK [fedora.linux_system_roles.storage : show storage_volumes] **************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:18 Thursday 21 July 2022 11:26:55 +0000 (0:00:00.043) 0:02:17.993 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_volumes": "VARIABLE IS NOT DEFINED!: 'storage_volumes' is undefined" } TASK [fedora.linux_system_roles.storage : get required packages] *************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 Thursday 21 July 2022 11:26:55 +0000 (0:00:00.042) 0:02:18.036 ********* ok: [/cache/rhel-9.qcow2.snap] => { "actions": [], "changed": false, "crypts": [], "leaves": [], "mounts": [], "packages": [], "pools": [], "volumes": [] } TASK [fedora.linux_system_roles.storage : enable copr repositories if needed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:35 Thursday 21 July 2022 11:26:57 +0000 (0:00:02.657) 0:02:20.693 ********* included: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml for /cache/rhel-9.qcow2.snap TASK [fedora.linux_system_roles.storage : check if the COPR support packages should be installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:2 Thursday 21 July 2022 11:26:57 +0000 (0:00:00.053) 0:02:20.746 ********* TASK [fedora.linux_system_roles.storage : make sure COPR support packages are present] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:13 Thursday 21 July 2022 11:26:57 +0000 (0:00:00.036) 0:02:20.782 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : enable COPRs] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/enable_coprs.yml:18 Thursday 21 July 2022 11:26:57 +0000 (0:00:00.040) 0:02:20.822 ********* TASK [fedora.linux_system_roles.storage : make sure required packages are installed] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:41 Thursday 21 July 2022 11:26:57 +0000 (0:00:00.037) 0:02:20.859 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [fedora.linux_system_roles.storage : get service facts] ******************* task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Thursday 21 July 2022 11:26:58 +0000 (0:00:00.840) 0:02:21.700 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "services": { "NetworkManager-dispatcher.service": { "name": "NetworkManager-dispatcher.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "NetworkManager-wait-online.service": { "name": "NetworkManager-wait-online.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "NetworkManager.service": { "name": "NetworkManager.service", "source": "systemd", "state": "running", "status": "enabled" }, "auditd.service": { "name": "auditd.service", "source": "systemd", "state": "running", "status": "enabled" }, "auth-rpcgss-module.service": { "name": "auth-rpcgss-module.service", "source": "systemd", "state": "stopped", "status": "static" }, "autofs.service": { "name": "autofs.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "autovt@.service": { "name": "autovt@.service", "source": "systemd", "state": "unknown", "status": "alias" }, "blivet.service": { "name": "blivet.service", "source": "systemd", "state": "inactive", "status": "static" }, "blk-availability.service": { "name": "blk-availability.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "chrony-wait.service": { "name": "chrony-wait.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "chronyd.service": { "name": "chronyd.service", "source": "systemd", "state": "running", "status": "enabled" }, "cloud-config.service": { "name": "cloud-config.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-final.service": { "name": "cloud-final.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init-local.service": { "name": "cloud-init-local.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cloud-init.service": { "name": "cloud-init.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "cockpit-motd.service": { "name": "cockpit-motd.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-http.service": { "name": "cockpit-wsinstance-http.service", "source": "systemd", "state": "inactive", "status": "static" }, "cockpit-wsinstance-https-factory@.service": { "name": "cockpit-wsinstance-https-factory@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit-wsinstance-https@.service": { "name": "cockpit-wsinstance-https@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cockpit.service": { "name": "cockpit.service", "source": "systemd", "state": "inactive", "status": "static" }, "console-getty.service": { "name": "console-getty.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "container-getty@.service": { "name": "container-getty@.service", "source": "systemd", "state": "unknown", "status": "static" }, "cpupower.service": { "name": "cpupower.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "crond.service": { "name": "crond.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-broker.service": { "name": "dbus-broker.service", "source": "systemd", "state": "running", "status": "enabled" }, "dbus-org.freedesktop.hostname1.service": { "name": "dbus-org.freedesktop.hostname1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.locale1.service": { "name": "dbus-org.freedesktop.locale1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.login1.service": { "name": "dbus-org.freedesktop.login1.service", "source": "systemd", "state": "active", "status": "alias" }, "dbus-org.freedesktop.nm-dispatcher.service": { "name": "dbus-org.freedesktop.nm-dispatcher.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus-org.freedesktop.timedate1.service": { "name": "dbus-org.freedesktop.timedate1.service", "source": "systemd", "state": "inactive", "status": "alias" }, "dbus.service": { "name": "dbus.service", "source": "systemd", "state": "active", "status": "alias" }, "debug-shell.service": { "name": "debug-shell.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "display-manager.service": { "name": "display-manager.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "dm-event.service": { "name": "dm-event.service", "source": "systemd", "state": "running", "status": "static" }, "dnf-makecache.service": { "name": "dnf-makecache.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-cmdline.service": { "name": "dracut-cmdline.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-initqueue.service": { "name": "dracut-initqueue.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-mount.service": { "name": "dracut-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-mount.service": { "name": "dracut-pre-mount.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-pivot.service": { "name": "dracut-pre-pivot.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-trigger.service": { "name": "dracut-pre-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-pre-udev.service": { "name": "dracut-pre-udev.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown-onfailure.service": { "name": "dracut-shutdown-onfailure.service", "source": "systemd", "state": "stopped", "status": "static" }, "dracut-shutdown.service": { "name": "dracut-shutdown.service", "source": "systemd", "state": "stopped", "status": "static" }, "emergency.service": { "name": "emergency.service", "source": "systemd", "state": "stopped", "status": "static" }, "fcoe.service": { "name": "fcoe.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "fstrim.service": { "name": "fstrim.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-offline-update.service": { "name": "fwupd-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd-refresh.service": { "name": "fwupd-refresh.service", "source": "systemd", "state": "inactive", "status": "static" }, "fwupd.service": { "name": "fwupd.service", "source": "systemd", "state": "inactive", "status": "static" }, "getty@.service": { "name": "getty@.service", "source": "systemd", "state": "unknown", "status": "enabled" }, "getty@tty1.service": { "name": "getty@tty1.service", "source": "systemd", "state": "running", "status": "active" }, "grub-boot-indeterminate.service": { "name": "grub-boot-indeterminate.service", "source": "systemd", "state": "inactive", "status": "static" }, "grub2-systemd-integration.service": { "name": "grub2-systemd-integration.service", "source": "systemd", "state": "inactive", "status": "static" }, "gssproxy.service": { "name": "gssproxy.service", "source": "systemd", "state": "running", "status": "disabled" }, "initrd-cleanup.service": { "name": "initrd-cleanup.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-parse-etc.service": { "name": "initrd-parse-etc.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-switch-root.service": { "name": "initrd-switch-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "initrd-udevadm-cleanup-db.service": { "name": "initrd-udevadm-cleanup-db.service", "source": "systemd", "state": "stopped", "status": "static" }, "insights-client-boot.service": { "name": "insights-client-boot.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "insights-client-results.service": { "name": "insights-client-results.service", "source": "systemd", "state": "inactive", "status": "static" }, "insights-client.service": { "name": "insights-client.service", "source": "systemd", "state": "inactive", "status": "static" }, "irqbalance.service": { "name": "irqbalance.service", "source": "systemd", "state": "running", "status": "enabled" }, "iscsi-shutdown.service": { "name": "iscsi-shutdown.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsi.service": { "name": "iscsi.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "iscsid.service": { "name": "iscsid.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "kdump.service": { "name": "kdump.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "kmod-static-nodes.service": { "name": "kmod-static-nodes.service", "source": "systemd", "state": "stopped", "status": "static" }, "kvm_stat.service": { "name": "kvm_stat.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "ldconfig.service": { "name": "ldconfig.service", "source": "systemd", "state": "stopped", "status": "static" }, "logrotate.service": { "name": "logrotate.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-activation-early.service": { "name": "lvm2-activation-early.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "lvm2-lvmpolld.service": { "name": "lvm2-lvmpolld.service", "source": "systemd", "state": "stopped", "status": "static" }, "lvm2-monitor.service": { "name": "lvm2-monitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "man-db-cache-update.service": { "name": "man-db-cache-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "man-db-restart-cache-update.service": { "name": "man-db-restart-cache-update.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "mdadm-grow-continue@.service": { "name": "mdadm-grow-continue@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdadm-last-resort@.service": { "name": "mdadm-last-resort@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdcheck_continue.service": { "name": "mdcheck_continue.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdcheck_start.service": { "name": "mdcheck_start.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmon@.service": { "name": "mdmon@.service", "source": "systemd", "state": "unknown", "status": "static" }, "mdmonitor-oneshot.service": { "name": "mdmonitor-oneshot.service", "source": "systemd", "state": "inactive", "status": "static" }, "mdmonitor.service": { "name": "mdmonitor.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "microcode.service": { "name": "microcode.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "modprobe@.service": { "name": "modprobe@.service", "source": "systemd", "state": "unknown", "status": "static" }, "modprobe@configfs.service": { "name": "modprobe@configfs.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@drm.service": { "name": "modprobe@drm.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "modprobe@fuse.service": { "name": "modprobe@fuse.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "multipathd.service": { "name": "multipathd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "ndctl-monitor.service": { "name": "ndctl-monitor.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "network.service": { "name": "network.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "nfs-blkmap.service": { "name": "nfs-blkmap.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "nfs-idmapd.service": { "name": "nfs-idmapd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-mountd.service": { "name": "nfs-mountd.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfs-server.service": { "name": "nfs-server.service", "source": "systemd", "state": "stopped", "status": "disabled" }, "nfs-utils.service": { "name": "nfs-utils.service", "source": "systemd", "state": "stopped", "status": "static" }, "nfsdcld.service": { "name": "nfsdcld.service", "source": "systemd", "state": "stopped", "status": "static" }, "nis-domainname.service": { "name": "nis-domainname.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "nm-priv-helper.service": { "name": "nm-priv-helper.service", "source": "systemd", "state": "inactive", "status": "static" }, "ntpd.service": { "name": "ntpd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ntpdate.service": { "name": "ntpdate.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "oddjobd.service": { "name": "oddjobd.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "packagekit-offline-update.service": { "name": "packagekit-offline-update.service", "source": "systemd", "state": "inactive", "status": "static" }, "packagekit.service": { "name": "packagekit.service", "source": "systemd", "state": "inactive", "status": "static" }, "pam_namespace.service": { "name": "pam_namespace.service", "source": "systemd", "state": "inactive", "status": "static" }, "plymouth-quit-wait.service": { "name": "plymouth-quit-wait.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "plymouth-start.service": { "name": "plymouth-start.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "polkit.service": { "name": "polkit.service", "source": "systemd", "state": "running", "status": "static" }, "qemu-guest-agent.service": { "name": "qemu-guest-agent.service", "source": "systemd", "state": "inactive", "status": "enabled" }, "quotaon.service": { "name": "quotaon.service", "source": "systemd", "state": "inactive", "status": "static" }, "raid-check.service": { "name": "raid-check.service", "source": "systemd", "state": "inactive", "status": "static" }, "rbdmap.service": { "name": "rbdmap.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rc-local.service": { "name": "rc-local.service", "source": "systemd", "state": "stopped", "status": "static" }, "rdisc.service": { "name": "rdisc.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rescue.service": { "name": "rescue.service", "source": "systemd", "state": "stopped", "status": "static" }, "rhsm-facts.service": { "name": "rhsm-facts.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsm.service": { "name": "rhsm.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rhsmcertd.service": { "name": "rhsmcertd.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpc-gssd.service": { "name": "rpc-gssd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd-notify.service": { "name": "rpc-statd-notify.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-statd.service": { "name": "rpc-statd.service", "source": "systemd", "state": "stopped", "status": "static" }, "rpc-svcgssd.service": { "name": "rpc-svcgssd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "rpcbind.service": { "name": "rpcbind.service", "source": "systemd", "state": "running", "status": "enabled" }, "rpmdb-rebuild.service": { "name": "rpmdb-rebuild.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "rsyslog.service": { "name": "rsyslog.service", "source": "systemd", "state": "running", "status": "enabled" }, "selinux-autorelabel-mark.service": { "name": "selinux-autorelabel-mark.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "selinux-autorelabel.service": { "name": "selinux-autorelabel.service", "source": "systemd", "state": "inactive", "status": "static" }, "serial-getty@.service": { "name": "serial-getty@.service", "source": "systemd", "state": "unknown", "status": "indirect" }, "serial-getty@ttyS0.service": { "name": "serial-getty@ttyS0.service", "source": "systemd", "state": "running", "status": "active" }, "sntp.service": { "name": "sntp.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen.service": { "name": "sshd-keygen.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "sshd-keygen@.service": { "name": "sshd-keygen@.service", "source": "systemd", "state": "unknown", "status": "disabled" }, "sshd-keygen@ecdsa.service": { "name": "sshd-keygen@ecdsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@ed25519.service": { "name": "sshd-keygen@ed25519.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd-keygen@rsa.service": { "name": "sshd-keygen@rsa.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "sshd.service": { "name": "sshd.service", "source": "systemd", "state": "running", "status": "enabled" }, "sshd@.service": { "name": "sshd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "sssd-autofs.service": { "name": "sssd-autofs.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-kcm.service": { "name": "sssd-kcm.service", "source": "systemd", "state": "stopped", "status": "indirect" }, "sssd-nss.service": { "name": "sssd-nss.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pac.service": { "name": "sssd-pac.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-pam.service": { "name": "sssd-pam.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-ssh.service": { "name": "sssd-ssh.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd-sudo.service": { "name": "sssd-sudo.service", "source": "systemd", "state": "inactive", "status": "indirect" }, "sssd.service": { "name": "sssd.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "syslog.service": { "name": "syslog.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "system-update-cleanup.service": { "name": "system-update-cleanup.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-ask-password-console.service": { "name": "systemd-ask-password-console.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-ask-password-wall.service": { "name": "systemd-ask-password-wall.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-backlight@.service": { "name": "systemd-backlight@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-binfmt.service": { "name": "systemd-binfmt.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-boot-check-no-failures.service": { "name": "systemd-boot-check-no-failures.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-coredump@.service": { "name": "systemd-coredump@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-exit.service": { "name": "systemd-exit.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-firstboot.service": { "name": "systemd-firstboot.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck-root.service": { "name": "systemd-fsck-root.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-fsck@.service": { "name": "systemd-fsck@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service": { "name": "systemd-fsck@dev-disk-by\\x2duuid-7B77\\x2d95E7.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vda2.service": { "name": "systemd-fsck@dev-vda2.service", "source": "systemd", "state": "stopped", "status": "inactive" }, "systemd-fsck@dev-vdb1.service": { "name": "systemd-fsck@dev-vdb1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-fsck@dev-vdc1.service": { "name": "systemd-fsck@dev-vdc1.service", "source": "systemd", "state": "stopped", "status": "active" }, "systemd-halt.service": { "name": "systemd-halt.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hibernate-resume@.service": { "name": "systemd-hibernate-resume@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-hibernate.service": { "name": "systemd-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hostnamed.service": { "name": "systemd-hostnamed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-hwdb-update.service": { "name": "systemd-hwdb-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-hybrid-sleep.service": { "name": "systemd-hybrid-sleep.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-initctl.service": { "name": "systemd-initctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-catalog-update.service": { "name": "systemd-journal-catalog-update.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journal-flush.service": { "name": "systemd-journal-flush.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-journald.service": { "name": "systemd-journald.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-journald@.service": { "name": "systemd-journald@.service", "source": "systemd", "state": "unknown", "status": "static" }, "systemd-kexec.service": { "name": "systemd-kexec.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-localed.service": { "name": "systemd-localed.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-logind.service": { "name": "systemd-logind.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-machine-id-commit.service": { "name": "systemd-machine-id-commit.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-modules-load.service": { "name": "systemd-modules-load.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-network-generator.service": { "name": "systemd-network-generator.service", "source": "systemd", "state": "stopped", "status": "enabled" }, "systemd-poweroff.service": { "name": "systemd-poweroff.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-pstore.service": { "name": "systemd-pstore.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-quotacheck.service": { "name": "systemd-quotacheck.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-random-seed.service": { "name": "systemd-random-seed.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-reboot.service": { "name": "systemd-reboot.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-remount-fs.service": { "name": "systemd-remount-fs.service", "source": "systemd", "state": "stopped", "status": "enabled-runtime" }, "systemd-repart.service": { "name": "systemd-repart.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-rfkill.service": { "name": "systemd-rfkill.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-suspend-then-hibernate.service": { "name": "systemd-suspend-then-hibernate.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-suspend.service": { "name": "systemd-suspend.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-sysctl.service": { "name": "systemd-sysctl.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-sysext.service": { "name": "systemd-sysext.service", "source": "systemd", "state": "inactive", "status": "disabled" }, "systemd-sysusers.service": { "name": "systemd-sysusers.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-timedated.service": { "name": "systemd-timedated.service", "source": "systemd", "state": "inactive", "status": "static" }, "systemd-timesyncd.service": { "name": "systemd-timesyncd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "systemd-tmpfiles-clean.service": { "name": "systemd-tmpfiles-clean.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup-dev.service": { "name": "systemd-tmpfiles-setup-dev.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-tmpfiles-setup.service": { "name": "systemd-tmpfiles-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-settle.service": { "name": "systemd-udev-settle.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udev-trigger.service": { "name": "systemd-udev-trigger.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-udevd.service": { "name": "systemd-udevd.service", "source": "systemd", "state": "running", "status": "static" }, "systemd-update-done.service": { "name": "systemd-update-done.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp-runlevel.service": { "name": "systemd-update-utmp-runlevel.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-update-utmp.service": { "name": "systemd-update-utmp.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-user-sessions.service": { "name": "systemd-user-sessions.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-vconsole-setup.service": { "name": "systemd-vconsole-setup.service", "source": "systemd", "state": "stopped", "status": "static" }, "systemd-volatile-root.service": { "name": "systemd-volatile-root.service", "source": "systemd", "state": "inactive", "status": "static" }, "teamd@.service": { "name": "teamd@.service", "source": "systemd", "state": "unknown", "status": "static" }, "tuned.service": { "name": "tuned.service", "source": "systemd", "state": "running", "status": "enabled" }, "user-runtime-dir@.service": { "name": "user-runtime-dir@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user-runtime-dir@0.service": { "name": "user-runtime-dir@0.service", "source": "systemd", "state": "stopped", "status": "active" }, "user@.service": { "name": "user@.service", "source": "systemd", "state": "unknown", "status": "static" }, "user@0.service": { "name": "user@0.service", "source": "systemd", "state": "running", "status": "active" }, "ypbind.service": { "name": "ypbind.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "yppasswdd.service": { "name": "yppasswdd.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypserv.service": { "name": "ypserv.service", "source": "systemd", "state": "stopped", "status": "not-found" }, "ypxfrd.service": { "name": "ypxfrd.service", "source": "systemd", "state": "stopped", "status": "not-found" } } }, "changed": false } TASK [fedora.linux_system_roles.storage : Set storage_cryptsetup_services] ***** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:53 Thursday 21 July 2022 11:27:00 +0000 (0:00:01.769) 0:02:23.470 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_cryptsetup_services": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : Mask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:58 Thursday 21 July 2022 11:27:00 +0000 (0:00:00.064) 0:02:23.534 ********* TASK [fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 Thursday 21 July 2022 11:27:00 +0000 (0:00:00.024) 0:02:23.558 ********* changed: [/cache/rhel-9.qcow2.snap] => { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "destroy device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "destroy device", "device": "/dev/vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdb1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdc1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc", "fs_type": "disklabel" } ], "changed": true, "crypts": [], "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/sdb", "/dev/sda", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [], "packages": [ "e2fsprogs", "xfsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } TASK [fedora.linux_system_roles.storage : Workaround for udev issue on some platforms] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:78 Thursday 21 July 2022 11:27:04 +0000 (0:00:04.174) 0:02:27.733 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : Unmask the systemd cryptsetup services] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:90 Thursday 21 July 2022 11:27:04 +0000 (0:00:00.043) 0:02:27.776 ********* TASK [fedora.linux_system_roles.storage : show blivet_output] ****************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:96 Thursday 21 July 2022 11:27:04 +0000 (0:00:00.025) 0:02:27.802 ********* ok: [/cache/rhel-9.qcow2.snap] => { "blivet_output": { "actions": [ { "action": "destroy format", "device": "/dev/mapper/vg1-lv1", "fs_type": "xfs" }, { "action": "destroy device", "device": "/dev/mapper/vg1-lv1", "fs_type": null }, { "action": "destroy device", "device": "/dev/mapper/vg1-tpool1", "fs_type": null }, { "action": "destroy device", "device": "/dev/vg1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sda1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sda", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sdb1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdb1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdb", "fs_type": "disklabel" }, { "action": "destroy format", "device": "/dev/sdc1", "fs_type": "lvmpv" }, { "action": "destroy device", "device": "/dev/sdc1", "fs_type": null }, { "action": "destroy format", "device": "/dev/sdc", "fs_type": "disklabel" } ], "changed": true, "crypts": [], "failed": false, "leaves": [ "/dev/sr0", "/dev/vda1", "/dev/vda2", "/dev/vda3", "/dev/vda4", "/dev/sdb", "/dev/sda", "/dev/sdc", "/dev/nvme0n1", "/dev/nvme1n1", "/dev/nvme2n1", "/dev/vdb1", "/dev/vdc1", "/dev/vdd", "/dev/vde", "/dev/vdf" ], "mounts": [], "packages": [ "e2fsprogs", "xfsprogs", "dosfstools" ], "pools": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ], "volumes": [] } } TASK [fedora.linux_system_roles.storage : set the list of pools for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:101 Thursday 21 July 2022 11:27:04 +0000 (0:00:00.045) 0:02:27.847 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] }, "changed": false } TASK [fedora.linux_system_roles.storage : set the list of volumes for test verification] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:105 Thursday 21 July 2022 11:27:04 +0000 (0:00:00.043) 0:02:27.890 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_volumes_list": [] }, "changed": false } TASK [fedora.linux_system_roles.storage : remove obsolete mounts] ************** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:121 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.042) 0:02:27.933 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:132 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.039) 0:02:27.973 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : set up new/current mounts] *********** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:137 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.026) 0:02:27.999 ********* TASK [fedora.linux_system_roles.storage : tell systemd to refresh its view of /etc/fstab] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:148 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.043) 0:02:28.043 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [fedora.linux_system_roles.storage : retrieve facts for the /etc/crypttab file] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:156 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.027) 0:02:28.070 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [fedora.linux_system_roles.storage : manage /etc/crypttab to account for changes we just made] *** task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:161 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.371) 0:02:28.442 ********* TASK [fedora.linux_system_roles.storage : Update facts] ************************ task path: /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:183 Thursday 21 July 2022 11:27:05 +0000 (0:00:00.025) 0:02:28.467 ********* ok: [/cache/rhel-9.qcow2.snap] META: role_complete for /cache/rhel-9.qcow2.snap TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/tests_create_thinp_then_remove.yml:125 Thursday 21 July 2022 11:27:06 +0000 (0:00:00.949) 0:02:29.417 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml for /cache/rhel-9.qcow2.snap TASK [Print out pool information] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:1 Thursday 21 July 2022 11:27:06 +0000 (0:00:00.057) 0:02:29.474 ********* ok: [/cache/rhel-9.qcow2.snap] => { "_storage_pools_list": [ { "disks": [ "sda", "sdb", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "name": "vg1", "raid_chunk_size": null, "raid_device_count": null, "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "state": "absent", "type": "lvm", "volumes": [ { "_device": "/dev/mapper/vg1-lv1", "_mount_id": "/dev/mapper/vg1-lv1", "_raw_device": "/dev/mapper/vg1-lv1", "cache_devices": [], "cache_mode": null, "cache_size": 0, "cached": false, "compression": null, "deduplication": null, "disks": [ "sdb", "sda", "sdc" ], "encryption": false, "encryption_cipher": null, "encryption_key": null, "encryption_key_size": null, "encryption_luks_version": null, "encryption_password": null, "fs_create_options": "", "fs_label": "", "fs_overwrite_existing": true, "fs_type": "xfs", "mount_check": 0, "mount_device_identifier": "uuid", "mount_options": "defaults", "mount_passno": 0, "mount_point": "/opt/test1", "name": "lv1", "raid_chunk_size": null, "raid_device_count": null, "raid_disks": [], "raid_level": null, "raid_metadata_version": null, "raid_spare_count": null, "size": "3g", "state": "present", "thin": true, "thin_pool_name": "tpool1", "thin_pool_size": "10g", "type": "lvm", "vdo_pool_size": null } ] } ] } TASK [Print out volume information] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:6 Thursday 21 July 2022 11:27:06 +0000 (0:00:00.102) 0:02:29.577 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Collect info about the volumes.] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:14 Thursday 21 July 2022 11:27:06 +0000 (0:00:00.081) 0:02:29.659 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "info": { "/dev/nvme0n1": { "fstype": "", "label": "", "name": "/dev/nvme0n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme1n1": { "fstype": "", "label": "", "name": "/dev/nvme1n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/nvme2n1": { "fstype": "", "label": "", "name": "/dev/nvme2n1", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sda": { "fstype": "", "label": "", "name": "/dev/sda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdb": { "fstype": "", "label": "", "name": "/dev/sdb", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sdc": { "fstype": "", "label": "", "name": "/dev/sdc", "size": "10G", "type": "disk", "uuid": "" }, "/dev/sr0": { "fstype": "iso9660", "label": "cidata", "name": "/dev/sr0", "size": "364K", "type": "rom", "uuid": "2022-07-21-11-24-24-00" }, "/dev/vda": { "fstype": "", "label": "", "name": "/dev/vda", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vda1": { "fstype": "", "label": "", "name": "/dev/vda1", "size": "1M", "type": "partition", "uuid": "" }, "/dev/vda2": { "fstype": "vfat", "label": "", "name": "/dev/vda2", "size": "200M", "type": "partition", "uuid": "7B77-95E7" }, "/dev/vda3": { "fstype": "xfs", "label": "boot", "name": "/dev/vda3", "size": "500M", "type": "partition", "uuid": "6e74e171-0370-451f-8340-f16ad2839183" }, "/dev/vda4": { "fstype": "xfs", "label": "root", "name": "/dev/vda4", "size": "9.3G", "type": "partition", "uuid": "1bb53e4d-984c-4316-908a-59b5a62fa30e" }, "/dev/vdb": { "fstype": "", "label": "", "name": "/dev/vdb", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdb1": { "fstype": "ext4", "label": "yumcache", "name": "/dev/vdb1", "size": "2G", "type": "partition", "uuid": "12be66f7-7f03-4d54-9a48-a520825d8fa9" }, "/dev/vdc": { "fstype": "", "label": "", "name": "/dev/vdc", "size": "2G", "type": "disk", "uuid": "" }, "/dev/vdc1": { "fstype": "ext4", "label": "yumvarlib", "name": "/dev/vdc1", "size": "2G", "type": "partition", "uuid": "9bd037ed-b1bb-4b9c-a008-9f07d34ab071" }, "/dev/vdd": { "fstype": "", "label": "", "name": "/dev/vdd", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vde": { "fstype": "", "label": "", "name": "/dev/vde", "size": "10G", "type": "disk", "uuid": "" }, "/dev/vdf": { "fstype": "", "label": "", "name": "/dev/vdf", "size": "10G", "type": "disk", "uuid": "" } } } TASK [Read the /etc/fstab file for volume existence] *************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:19 Thursday 21 July 2022 11:27:07 +0000 (0:00:00.379) 0:02:30.038 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/fstab" ], "delta": "0:00:01.003851", "end": "2022-07-21 07:27:08.108301", "rc": 0, "start": "2022-07-21 07:27:07.104450" } STDOUT: UUID=6e74e171-0370-451f-8340-f16ad2839183 /boot xfs defaults 0 0 UUID=1bb53e4d-984c-4316-908a-59b5a62fa30e / xfs defaults 0 0 UUID=7B77-95E7 /boot/efi vfat defaults,uid=0,gid=0,umask=077,shortname=winnt 0 2 /dev/vdb1 /var/cache/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 /dev/vdc1 /var/lib/dnf auto defaults,nofail,x-systemd.requires=cloud-init.service,_netdev,comment=cloudconfig 0 2 TASK [Read the /etc/crypttab file] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 Thursday 21 July 2022 11:27:08 +0000 (0:00:01.371) 0:02:31.410 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "cmd": [ "cat", "/etc/crypttab" ], "delta": "0:00:00.002446", "end": "2022-07-21 07:27:08.471812", "failed_when_result": false, "rc": 1, "start": "2022-07-21 07:27:08.469366" } STDERR: cat: /etc/crypttab: No such file or directory MSG: non-zero return code TASK [Verify the volumes listed in storage_pools were correctly managed] ******* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:33 Thursday 21 July 2022 11:27:08 +0000 (0:00:00.355) 0:02:31.766 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml for /cache/rhel-9.qcow2.snap => (item={'disks': ['sda', 'sdb', 'sdc'], 'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'name': 'vg1', 'raid_level': None, 'raid_device_count': None, 'raid_spare_count': None, 'raid_metadata_version': None, 'state': 'absent', 'type': 'lvm', 'volumes': [{'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}], 'raid_chunk_size': None}) TASK [Set _storage_pool_tests] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:5 Thursday 21 July 2022 11:27:08 +0000 (0:00:00.056) 0:02:31.822 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_pool_tests": [ "members", "volumes" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool.yml:18 Thursday 21 July 2022 11:27:08 +0000 (0:00:00.033) 0:02:31.856 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml for /cache/rhel-9.qcow2.snap => (item=members) included: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml for /cache/rhel-9.qcow2.snap => (item=volumes) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:1 Thursday 21 July 2022 11:27:08 +0000 (0:00:00.044) 0:02:31.900 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_count": "0", "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [Get the canonical device path for each member device] ******************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:6 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.050) 0:02:31.951 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:15 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.021) 0:02:31.972 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": "0" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:19 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.052) 0:02:32.025 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_pool_pvs": [] }, "changed": false } TASK [Verify PV count] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:23 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.050) 0:02:32.075 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:29 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.047) 0:02:32.122 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "disk" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:33 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.038) 0:02:32.160 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_pv_type": "partition" }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:37 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.051) 0:02:32.212 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check the type of each PV] *********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:41 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.025) 0:02:32.238 ********* TASK [Check MD RAID] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:50 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.021) 0:02:32.259 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml for /cache/rhel-9.qcow2.snap TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:6 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.041) 0:02:32.301 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:12 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.026) 0:02:32.327 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:16 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.025) 0:02:32.353 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:20 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.024) 0:02:32.378 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:24 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.023) 0:02:32.401 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:30 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.026) 0:02:32.427 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:36 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.025) 0:02:32.452 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-md.yml:44 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.033) 0:02:32.486 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_md_active_devices_re": null, "storage_test_md_metadata_version_re": null, "storage_test_md_spare_devices_re": null }, "changed": false } TASK [Check LVM RAID] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:53 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.095) 0:02:32.581 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member LVM RAID settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-lvmraid.yml:1 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.046) 0:02:32.627 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [Get information about LVM RAID] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:3 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.086) 0:02:32.713 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is LVM RAID] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:8 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.028) 0:02:32.742 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-lvmraid.yml:12 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.026) 0:02:32.769 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check Thin Pools] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:56 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.026) 0:02:32.795 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member thinpool settings] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-thin.yml:1 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.044) 0:02:32.840 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [Get information about thinpool] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:3 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.044) 0:02:32.884 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in correct thinpool (when thinp name is provided)] *** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:8 Thursday 21 July 2022 11:27:09 +0000 (0:00:00.026) 0:02:32.910 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check that volume is in thinpool (when thinp name is not provided)] ****** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:13 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.024) 0:02:32.935 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-thin.yml:17 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.024) 0:02:32.959 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check member encryption] ************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:59 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.023) 0:02:32.983 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml for /cache/rhel-9.qcow2.snap TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:4 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.046) 0:02:33.029 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Validate pool member LUKS settings] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:8 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.051) 0:02:33.080 ********* TASK [Validate pool member crypttab entries] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:15 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.023) 0:02:33.104 ********* TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-encryption.yml:22 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.023) 0:02:33.127 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_crypttab_key_file": null }, "changed": false } TASK [Check VDO] *************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:62 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.038) 0:02:33.165 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml for /cache/rhel-9.qcow2.snap TASK [Validate pool member VDO settings] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-members-vdo.yml:1 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.051) 0:02:33.216 ********* included: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [get information about VDO deduplication] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:3 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.048) 0:02:33.264 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:8 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.027) 0:02:33.291 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:11 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.027) 0:02:33.319 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:16 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.025) 0:02:33.344 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:21 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.024) 0:02:33.369 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:24 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.023) 0:02:33.393 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:29 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.024) 0:02:33.417 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/verify-pool-member-vdo.yml:39 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.024) 0:02:33.442 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_vdo_status": null }, "changed": false } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-members.yml:65 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.035) 0:02:33.477 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "__pvs_lvm_len": null, "_storage_test_expected_pv_count": null, "_storage_test_expected_pv_type": null, "_storage_test_pool_pvs": [], "_storage_test_pool_pvs_lvm": [] }, "changed": false } TASK [verify the volumes] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-pool-volumes.yml:3 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.034) 0:02:33.512 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml for /cache/rhel-9.qcow2.snap => (item={'encryption': False, 'encryption_cipher': None, 'encryption_key': None, 'encryption_key_size': None, 'encryption_luks_version': None, 'encryption_password': None, 'fs_create_options': '', 'fs_label': '', 'fs_type': 'xfs', 'mount_options': 'defaults', 'mount_point': '/opt/test1', 'name': 'lv1', 'raid_level': None, 'size': '3g', 'state': 'present', 'type': 'lvm', 'cached': False, 'cache_devices': [], 'cache_mode': None, 'cache_size': 0, 'compression': None, 'deduplication': None, 'raid_disks': [], 'thin_pool_name': 'tpool1', 'thin_pool_size': '10g', 'thin': True, 'vdo_pool_size': None, 'disks': ['sdb', 'sda', 'sdc'], 'fs_overwrite_existing': True, 'mount_check': 0, 'mount_passno': 0, 'mount_device_identifier': 'uuid', 'raid_device_count': None, 'raid_spare_count': None, 'raid_chunk_size': None, 'raid_metadata_version': None, '_device': '/dev/mapper/vg1-lv1', '_raw_device': '/dev/mapper/vg1-lv1', '_mount_id': '/dev/mapper/vg1-lv1'}) TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:2 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.041) 0:02:33.554 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": false, "_storage_volume_tests": [ "mount", "fstab", "fs", "device", "encryption", "md", "size", "cache" ] }, "changed": false } TASK [include_tasks] *********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:10 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.046) 0:02:33.600 ********* included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml for /cache/rhel-9.qcow2.snap => (item=mount) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml for /cache/rhel-9.qcow2.snap => (item=fstab) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml for /cache/rhel-9.qcow2.snap => (item=fs) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml for /cache/rhel-9.qcow2.snap => (item=device) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml for /cache/rhel-9.qcow2.snap => (item=encryption) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml for /cache/rhel-9.qcow2.snap => (item=md) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml for /cache/rhel-9.qcow2.snap => (item=size) included: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml for /cache/rhel-9.qcow2.snap => (item=cache) TASK [Get expected mount device based on device type] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:6 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.071) 0:02:33.672 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_device_path": "/dev/mapper/vg1-lv1" }, "changed": false } TASK [Set some facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:10 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.043) 0:02:33.715 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": [], "storage_test_mount_expected_match_count": "0", "storage_test_mount_point_matches": [], "storage_test_swap_expected_matches": "0" }, "changed": false } TASK [Verify the current mount state by device] ******************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:20 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.059) 0:02:33.775 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the current mount state by mount point] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:29 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.024) 0:02:33.800 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify the mount fs type] ************************************************ task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:37 Thursday 21 July 2022 11:27:10 +0000 (0:00:00.082) 0:02:33.883 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [command] ***************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:46 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.075) 0:02:33.958 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Gather swap info] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:50 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.026) 0:02:33.985 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify swap status] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:55 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.024) 0:02:34.010 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Unset facts] ************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-mount.yml:65 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.023) 0:02:34.033 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_mount_device_matches": null, "storage_test_mount_expected_match_count": null, "storage_test_mount_point_matches": null, "storage_test_swap_expected_matches": null, "storage_test_swaps": null, "storage_test_sys_node": null }, "changed": false } TASK [Set some variables for fstab checking] *********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:2 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.036) 0:02:34.069 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": "0", "storage_test_fstab_expected_mount_options_matches": "0", "storage_test_fstab_expected_mount_point_matches": "0", "storage_test_fstab_id_matches": [], "storage_test_fstab_mount_options_matches": [], "storage_test_fstab_mount_point_matches": [] }, "changed": false } TASK [Verify that the device identifier appears in /etc/fstab] ***************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:12 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.062) 0:02:34.132 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the fstab mount point] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:19 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.024) 0:02:34.156 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Verify mount_options] **************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:25 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.052) 0:02:34.208 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up variables] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fstab.yml:34 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.039) 0:02:34.248 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_fstab_expected_id_matches": null, "storage_test_fstab_expected_mount_options_matches": null, "storage_test_fstab_expected_mount_point_matches": null, "storage_test_fstab_id_matches": null, "storage_test_fstab_mount_options_matches": null, "storage_test_fstab_mount_point_matches": null }, "changed": false } TASK [Verify fs type] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:4 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.033) 0:02:34.281 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify fs label] ********************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-fs.yml:10 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.023) 0:02:34.305 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [See whether the device node is present] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:4 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.023) 0:02:34.328 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "stat": { "exists": false } } TASK [Verify the presence/absence of the device node] ************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:10 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.358) 0:02:34.686 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Make sure we got info about this volume] ********************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:15 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.038) 0:02:34.724 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [(1/2) Process volume type (set initial value)] *************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:21 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.026) 0:02:34.750 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "st_volume_type": "lvm" }, "changed": false } TASK [(2/2) Process volume type (get RAID value)] ****************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:25 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.042) 0:02:34.793 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the volume's device type] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-device.yml:30 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.026) 0:02:34.819 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Stat the LUKS device, if encrypted] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:3 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.025) 0:02:34.844 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Ensure cryptsetup is present] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:10 Thursday 21 July 2022 11:27:11 +0000 (0:00:00.025) 0:02:34.869 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false, "rc": 0, "results": [] } MSG: Nothing to do TASK [Collect LUKS info for this volume] *************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:15 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.821) 0:02:35.691 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the presence/absence of the LUKS device node] ********************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:21 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.026) 0:02:35.718 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify that the raw device is the same as the device if not encrypted] *** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:27 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.024) 0:02:35.742 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Make sure we got info about the LUKS volume if encrypted] **************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:33 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.023) 0:02:35.765 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Verify the LUKS volume's device type if encrypted] *********************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:39 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.023) 0:02:35.789 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS version] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:44 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.023) 0:02:35.813 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS key size] ***************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:50 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.025) 0:02:35.838 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check LUKS cipher] ******************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:56 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.025) 0:02:35.864 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:62 Thursday 21 July 2022 11:27:12 +0000 (0:00:00.024) 0:02:35.888 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": [], "_storage_test_expected_crypttab_entries": "0", "_storage_test_expected_crypttab_key_file": "-" }, "changed": false } TASK [Check for /etc/crypttab entry] ******************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:67 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.051) 0:02:35.940 ********* ok: [/cache/rhel-9.qcow2.snap] => { "changed": false } MSG: All assertions passed TASK [Validate the format of the crypttab entry] ******************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:72 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.050) 0:02:35.990 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check backing device of crypttab entry] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:78 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.037) 0:02:36.027 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check key file of crypttab entry] **************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:84 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.037) 0:02:36.065 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-encryption.yml:90 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.049) 0:02:36.114 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_crypttab_entries": null, "_storage_test_expected_crypttab_entries": null, "_storage_test_expected_crypttab_key_file": null }, "changed": false } TASK [get information about RAID] ********************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:7 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.044) 0:02:36.158 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:13 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.045) 0:02:36.204 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:17 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.041) 0:02:36.246 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:21 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.045) 0:02:36.291 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID active devices count] ***************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:25 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.039) 0:02:36.331 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID spare devices count] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:31 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.037) 0:02:36.368 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check RAID metadata version] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-md.yml:37 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.036) 0:02:36.404 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the actual size of the volume] ************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:3 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.038) 0:02:36.443 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested size of the volume] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:9 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.069) 0:02:36.513 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Establish base value for expected size] ********************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:15 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.040) 0:02:36.553 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:20 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.036) 0:02:36.589 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:25 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.035) 0:02:36.625 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:28 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.037) 0:02:36.663 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Get the size of parent/pool device] ************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:31 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.034) 0:02:36.697 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:36 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.033) 0:02:36.731 ********* skipping: [/cache/rhel-9.qcow2.snap] => {} TASK [Calculate the expected size based on pool size and percentage value] ***** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:39 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.033) 0:02:36.765 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:44 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.037) 0:02:36.802 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_actual_size": { "changed": false, "skip_reason": "Conditional result was False", "skipped": true } } TASK [debug] ******************************************************************* task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:47 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.037) 0:02:36.839 ********* ok: [/cache/rhel-9.qcow2.snap] => { "storage_test_expected_size": "4294967296" } TASK [assert] ****************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-size.yml:50 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.041) 0:02:36.881 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Get information about the LV] ******************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:6 Thursday 21 July 2022 11:27:13 +0000 (0:00:00.023) 0:02:36.905 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:14 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.026) 0:02:36.931 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [check segment type] ****************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:17 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.025) 0:02:36.957 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:22 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.025) 0:02:36.982 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [parse the requested cache size] ****************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:26 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.025) 0:02:37.007 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [set_fact] **************************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:32 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.027) 0:02:37.035 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Check cache size] ******************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume-cache.yml:36 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.025) 0:02:37.060 ********* skipping: [/cache/rhel-9.qcow2.snap] => { "changed": false, "skip_reason": "Conditional result was False" } TASK [Clean up facts] ********************************************************** task path: /tmp/tmpfc26zqih/tests/storage/test-verify-volume.yml:16 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.024) 0:02:37.085 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "_storage_test_volume_present": null }, "changed": false } TASK [Verify the volumes with no pool were correctly managed] ****************** task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:43 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.037) 0:02:37.123 ********* TASK [Clean up variable namespace] ********************************************* task path: /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:53 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.024) 0:02:37.147 ********* ok: [/cache/rhel-9.qcow2.snap] => { "ansible_facts": { "storage_test_blkinfo": null, "storage_test_crypttab": null, "storage_test_fstab": null }, "changed": false } META: ran handlers META: ran handlers PLAY RECAP ********************************************************************* /cache/rhel-9.qcow2.snap : ok=646 changed=8 unreachable=0 failed=0 skipped=476 rescued=0 ignored=0 Thursday 21 July 2022 11:27:14 +0000 (0:00:00.044) 0:02:37.192 ********* =============================================================================== fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 4.33s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 4.17s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 3.50s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 3.04s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : get service facts ------------------- 2.83s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.79s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : get required packages --------------- 2.77s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get required packages --------------- 2.75s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : manage the pools and volumes to match the specified state --- 2.69s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:64 fedora.linux_system_roles.storage : get required packages --------------- 2.66s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get required packages --------------- 2.66s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get required packages --------------- 2.54s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:23 fedora.linux_system_roles.storage : get service facts ------------------- 1.80s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.77s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.75s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.68s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.67s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 fedora.linux_system_roles.storage : get service facts ------------------- 1.63s /tmp/tmpi19f9hzy/ansible_collections/fedora/linux_system_roles/roles/storage/tasks/main-blivet.yml:47 Read the /etc/crypttab file --------------------------------------------- 1.38s /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 --------------------- Read the /etc/crypttab file --------------------------------------------- 1.37s /tmp/tmpfc26zqih/tests/storage/verify-role-results.yml:24 ---------------------